pte 77 arch/alpha/include/asm/mmzone.h #define pte_pfn(pte) (pte_val(pte) >> 32) pte 81 arch/alpha/include/asm/mmzone.h pte_t pte; \ pte 85 arch/alpha/include/asm/mmzone.h pte_val(pte) = pfn | pgprot_val(pgprot); \ pte 87 arch/alpha/include/asm/mmzone.h pte; \ pte 31 arch/alpha/include/asm/page.h typedef struct { unsigned long pte; } pte_t; pte 36 arch/alpha/include/asm/page.h #define pte_val(x) ((x).pte) pte 17 arch/alpha/include/asm/pgalloc.h pmd_populate(struct mm_struct *mm, pmd_t *pmd, pgtable_t pte) pte 19 arch/alpha/include/asm/pgalloc.h pmd_set(pmd, (pte_t *)(page_to_pa(pte) + PAGE_OFFSET)); pte 24 arch/alpha/include/asm/pgalloc.h pmd_populate_kernel(struct mm_struct *mm, pmd_t *pmd, pte_t *pte) pte 26 arch/alpha/include/asm/pgalloc.h pmd_set(pmd, pte); pte 209 arch/alpha/include/asm/pgtable.h #define pte_pfn(pte) (pte_val(pte) >> 32) pte 210 arch/alpha/include/asm/pgtable.h #define pte_page(pte) pfn_to_page(pte_pfn(pte)) pte 213 arch/alpha/include/asm/pgtable.h pte_t pte; \ pte 215 arch/alpha/include/asm/pgtable.h pte_val(pte) = (page_to_pfn(page) << 32) | pgprot_val(pgprot); \ pte 216 arch/alpha/include/asm/pgtable.h pte; \ pte 221 arch/alpha/include/asm/pgtable.h { pte_t pte; pte_val(pte) = (PHYS_TWIDDLE(physpfn) << 32) | pgprot_val(pgprot); return pte; } pte 223 arch/alpha/include/asm/pgtable.h extern inline pte_t pte_modify(pte_t pte, pgprot_t newprot) pte 224 arch/alpha/include/asm/pgtable.h { pte_val(pte) = (pte_val(pte) & _PAGE_CHG_MASK) | pgprot_val(newprot); return pte; } pte 247 arch/alpha/include/asm/pgtable.h extern inline int pte_none(pte_t pte) { return !pte_val(pte); } pte 248 arch/alpha/include/asm/pgtable.h extern inline int pte_present(pte_t pte) { return pte_val(pte) & _PAGE_VALID; } pte 268 arch/alpha/include/asm/pgtable.h extern inline int pte_write(pte_t pte) { return !(pte_val(pte) & _PAGE_FOW); } pte 269 arch/alpha/include/asm/pgtable.h extern inline int pte_dirty(pte_t pte) { return pte_val(pte) & _PAGE_DIRTY; } pte 270 arch/alpha/include/asm/pgtable.h extern inline int pte_young(pte_t pte) { return pte_val(pte) & _PAGE_ACCESSED; } pte 271 arch/alpha/include/asm/pgtable.h extern inline int pte_special(pte_t pte) { return 0; } pte 273 arch/alpha/include/asm/pgtable.h extern inline pte_t pte_wrprotect(pte_t pte) { pte_val(pte) |= _PAGE_FOW; return pte; } pte 274 arch/alpha/include/asm/pgtable.h extern inline pte_t pte_mkclean(pte_t pte) { pte_val(pte) &= ~(__DIRTY_BITS); return pte; } pte 275 arch/alpha/include/asm/pgtable.h extern inline pte_t pte_mkold(pte_t pte) { pte_val(pte) &= ~(__ACCESS_BITS); return pte; } pte 276 arch/alpha/include/asm/pgtable.h extern inline pte_t pte_mkwrite(pte_t pte) { pte_val(pte) &= ~_PAGE_FOW; return pte; } pte 277 arch/alpha/include/asm/pgtable.h extern inline pte_t pte_mkdirty(pte_t pte) { pte_val(pte) |= __DIRTY_BITS; return pte; } pte 278 arch/alpha/include/asm/pgtable.h extern inline pte_t pte_mkyoung(pte_t pte) { pte_val(pte) |= __ACCESS_BITS; return pte; } pte 279 arch/alpha/include/asm/pgtable.h extern inline pte_t pte_mkspecial(pte_t pte) { return pte; } pte 321 arch/alpha/include/asm/pgtable.h #define pte_unmap(pte) do { } while (0) pte 339 arch/alpha/include/asm/pgtable.h { pte_t pte; pte_val(pte) = (type << 32) | (offset << 40); return pte; } pte 344 arch/alpha/include/asm/pgtable.h #define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val(pte) }) pte 7 arch/alpha/include/asm/tlb.h #define __pte_free_tlb(tlb, pte, address) pte_free((tlb)->mm, pte) pte 330 arch/alpha/kernel/core_cia.c unsigned long *ppte, pte; pte 338 arch/alpha/kernel/core_cia.c pte = (virt_to_phys(ppte) >> (PAGE_SHIFT - 1)) | 1; pte 341 arch/alpha/kernel/core_cia.c ppte[i] = pte; pte 319 arch/alpha/kernel/core_irongate.c u32 *mmio_regs, *gatt_pages, *cur_gatt, pte; pte 371 arch/alpha/kernel/core_irongate.c pte = cur_gatt[GET_GATT_OFF(baddr)] & ~1; pte 373 arch/alpha/kernel/core_irongate.c cur_gatt, pte); pte 388 arch/alpha/kernel/core_irongate.c pte = cur_gatt[GET_GATT_OFF(baddr)] & ~1; pte 391 arch/alpha/kernel/core_irongate.c pte, PAGE_SIZE, 0)) { pte 1043 arch/alpha/kernel/core_marvel.c unsigned long pte; pte 1051 arch/alpha/kernel/core_marvel.c pte = aper->arena->ptes[baddr >> PAGE_SHIFT]; pte 1052 arch/alpha/kernel/core_marvel.c if (!(pte & 1)) { pte 1056 arch/alpha/kernel/core_marvel.c return (pte >> 1) << PAGE_SHIFT; pte 704 arch/alpha/kernel/core_titan.c unsigned long pte; pte 712 arch/alpha/kernel/core_titan.c pte = aper->arena->ptes[baddr >> PAGE_SHIFT]; pte 713 arch/alpha/kernel/core_titan.c if (!(pte & 1)) { pte 718 arch/alpha/kernel/core_titan.c return (pte >> 1) << PAGE_SHIFT; pte 237 arch/arc/include/asm/arcregs.h unsigned int pad:6, fbe:2, tqe:2, ts:4, ft:1, rse:2, pte:3, bce:3, ver:8; pte 239 arch/arc/include/asm/arcregs.h unsigned int ver:8, bce:3, pte:3, rse:2, ft:1, ts:4, tqe:2, fbe:2, pad:6; pte 19 arch/arc/include/asm/hugepage.h static inline pmd_t pte_pmd(pte_t pte) pte 21 arch/arc/include/asm/hugepage.h return __pmd(pte_val(pte)); pte 32 arch/arc/include/asm/page.h unsigned long long pte; pte 34 arch/arc/include/asm/page.h unsigned long pte; pte 44 arch/arc/include/asm/page.h #define pte_val(x) ((x).pte) pte 36 arch/arc/include/asm/pgalloc.h pmd_populate_kernel(struct mm_struct *mm, pmd_t *pmd, pte_t *pte) pte 38 arch/arc/include/asm/pgalloc.h pmd_set(pmd, pte); pte 92 arch/arc/include/asm/pgalloc.h pte_t *pte; pte 94 arch/arc/include/asm/pgalloc.h pte = (pte_t *) __get_free_pages(GFP_KERNEL | __GFP_ZERO, pte 97 arch/arc/include/asm/pgalloc.h return pte; pte 119 arch/arc/include/asm/pgalloc.h static inline void pte_free_kernel(struct mm_struct *mm, pte_t *pte) pte 121 arch/arc/include/asm/pgalloc.h free_pages((unsigned long)pte, __get_order_pte()); /* takes phy addr */ pte 130 arch/arc/include/asm/pgalloc.h #define __pte_free_tlb(tlb, pte, addr) pte_free((tlb)->mm, pte) pte 252 arch/arc/include/asm/pgtable.h #define pte_unmap(pte) do { } while (0) pte 253 arch/arc/include/asm/pgtable.h #define pte_unmap_nested(pte) do { } while (0) pte 279 arch/arc/include/asm/pgtable.h #define pte_page(pte) pfn_to_page(pte_pfn(pte)) pte 284 arch/arc/include/asm/pgtable.h #define pte_pfn(pte) (pte_val(pte) >> PAGE_SHIFT) pte 299 arch/arc/include/asm/pgtable.h #define pte_read(pte) (pte_val(pte) & _PAGE_READ) pte 300 arch/arc/include/asm/pgtable.h #define pte_write(pte) (pte_val(pte) & _PAGE_WRITE) pte 301 arch/arc/include/asm/pgtable.h #define pte_dirty(pte) (pte_val(pte) & _PAGE_DIRTY) pte 302 arch/arc/include/asm/pgtable.h #define pte_young(pte) (pte_val(pte) & _PAGE_ACCESSED) pte 303 arch/arc/include/asm/pgtable.h #define pte_special(pte) (pte_val(pte) & _PAGE_SPECIAL) pte 306 arch/arc/include/asm/pgtable.h static inline pte_t pte_##fn(pte_t pte) { pte_val(pte) op; return pte; } pte 320 arch/arc/include/asm/pgtable.h static inline pte_t pte_modify(pte_t pte, pgprot_t newprot) pte 322 arch/arc/include/asm/pgtable.h return __pte((pte_val(pte) & _PAGE_CHG_MASK) | pgprot_val(newprot)); pte 380 arch/arc/include/asm/pgtable.h #define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val(pte) }) pte 223 arch/arc/kernel/setup.c cpu->bpu.num_pred = 2048 << bpu.pte; pte 659 arch/arc/mm/tlb.c pte_t pte = __pte(pmd_val(*pmd)); pte 660 arch/arc/mm/tlb.c update_mmu_cache(vma, addr, &pte); pte 270 arch/arm/include/asm/kvm_host.h int kvm_set_spte_hva(struct kvm *kvm, unsigned long hva, pte_t pte); pte 134 arch/arm/include/asm/kvm_mmu.h static inline pte_t kvm_s2pte_mkwrite(pte_t pte) pte 136 arch/arm/include/asm/kvm_mmu.h pte_val(pte) |= L_PTE_S2_RDWR; pte 137 arch/arm/include/asm/kvm_mmu.h return pte; pte 146 arch/arm/include/asm/kvm_mmu.h static inline pte_t kvm_s2pte_mkexec(pte_t pte) pte 148 arch/arm/include/asm/kvm_mmu.h pte_val(pte) &= ~L_PTE_XN; pte 149 arch/arm/include/asm/kvm_mmu.h return pte; pte 158 arch/arm/include/asm/kvm_mmu.h static inline void kvm_set_s2pte_readonly(pte_t *pte) pte 160 arch/arm/include/asm/kvm_mmu.h pte_val(*pte) = (pte_val(*pte) & ~L_PTE_S2_RDWR) | L_PTE_S2_RDONLY; pte 163 arch/arm/include/asm/kvm_mmu.h static inline bool kvm_s2pte_readonly(pte_t *pte) pte 165 arch/arm/include/asm/kvm_mmu.h return (pte_val(*pte) & L_PTE_S2_RDWR) == L_PTE_S2_RDONLY; pte 168 arch/arm/include/asm/kvm_mmu.h static inline bool kvm_s2pte_exec(pte_t *pte) pte 170 arch/arm/include/asm/kvm_mmu.h return !(pte_val(*pte) & L_PTE_XN); pte 301 arch/arm/include/asm/kvm_mmu.h static inline void __kvm_flush_dcache_pte(pte_t pte) pte 303 arch/arm/include/asm/kvm_mmu.h void *va = kmap_atomic(pte_page(pte)); pte 48 arch/arm/include/asm/pgalloc.h #define pud_populate(mm,pmd,pte) BUG() pte 55 arch/arm/include/asm/pgalloc.h static inline void clean_pte_table(pte_t *pte) pte 57 arch/arm/include/asm/pgalloc.h clean_dcache_area(pte + PTE_HWTABLE_PTRS, PTE_HWTABLE_SIZE); pte 84 arch/arm/include/asm/pgalloc.h pte_t *pte = __pte_alloc_one_kernel(mm); pte 86 arch/arm/include/asm/pgalloc.h if (pte) pte 87 arch/arm/include/asm/pgalloc.h clean_pte_table(pte); pte 89 arch/arm/include/asm/pgalloc.h return pte; pte 101 arch/arm/include/asm/pgalloc.h struct page *pte; pte 103 arch/arm/include/asm/pgalloc.h pte = __pte_alloc_one(mm, GFP_PGTABLE_USER | PGTABLE_HIGHMEM); pte 104 arch/arm/include/asm/pgalloc.h if (!pte) pte 106 arch/arm/include/asm/pgalloc.h if (!PageHighMem(pte)) pte 107 arch/arm/include/asm/pgalloc.h clean_pte_table(page_address(pte)); pte 108 arch/arm/include/asm/pgalloc.h return pte; pte 111 arch/arm/include/asm/pgalloc.h static inline void __pmd_populate(pmd_t *pmdp, phys_addr_t pte, pte 114 arch/arm/include/asm/pgalloc.h pmdval_t pmdval = (pte + PTE_HWTABLE_OFF) | prot; pte 21 arch/arm/include/asm/pgtable-2level-types.h typedef struct { pteval_t pte; } pte_t; pte 26 arch/arm/include/asm/pgtable-2level-types.h #define pte_val(x) ((x).pte) pte 212 arch/arm/include/asm/pgtable-2level.h #define set_pte_ext(ptep,pte,ext) cpu_set_pte_ext(ptep,pte,ext) pte 213 arch/arm/include/asm/pgtable-2level.h #define pte_special(pte) (0) pte 214 arch/arm/include/asm/pgtable-2level.h static inline pte_t pte_mkspecial(pte_t pte) { return pte; } pte 24 arch/arm/include/asm/pgtable-3level-types.h typedef struct { pteval_t pte; } pte_t; pte 29 arch/arm/include/asm/pgtable-3level-types.h #define pte_val(x) ((x).pte) pte 193 arch/arm/include/asm/pgtable-3level.h #define set_pte_ext(ptep,pte,ext) cpu_set_pte_ext(ptep,__pte(pte_val(pte)|(ext))) pte 195 arch/arm/include/asm/pgtable-3level.h #define pte_huge(pte) (pte_val(pte) && !(pte_val(pte) & PTE_TABLE_BIT)) pte 196 arch/arm/include/asm/pgtable-3level.h #define pte_mkhuge(pte) (__pte(pte_val(pte) & ~PTE_TABLE_BIT)) pte 204 arch/arm/include/asm/pgtable-3level.h #define pte_special(pte) (pte_isset((pte), L_PTE_SPECIAL)) pte 205 arch/arm/include/asm/pgtable-3level.h static inline pte_t pte_mkspecial(pte_t pte) pte 207 arch/arm/include/asm/pgtable-3level.h pte_val(pte) |= L_PTE_SPECIAL; pte 208 arch/arm/include/asm/pgtable-3level.h return pte; pte 48 arch/arm/include/asm/pgtable-nommu.h #define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val(pte) }) pte 53 arch/arm/include/asm/pgtable.h #define pte_ERROR(pte) __pte_error(__FILE__, __LINE__, pte) pte 198 arch/arm/include/asm/pgtable.h #define __pte_unmap(pte) do { } while (0) pte 201 arch/arm/include/asm/pgtable.h #define __pte_unmap(pte) kunmap_atomic(pte) pte 209 arch/arm/include/asm/pgtable.h #define pte_unmap(pte) __pte_unmap(pte) pte 211 arch/arm/include/asm/pgtable.h #define pte_pfn(pte) ((pte_val(pte) & PHYS_MASK) >> PAGE_SHIFT) pte 214 arch/arm/include/asm/pgtable.h #define pte_page(pte) pfn_to_page(pte_pfn(pte)) pte 219 arch/arm/include/asm/pgtable.h #define pte_isset(pte, val) ((u32)(val) == (val) ? pte_val(pte) & (val) \ pte 220 arch/arm/include/asm/pgtable.h : !!(pte_val(pte) & (val))) pte 221 arch/arm/include/asm/pgtable.h #define pte_isclear(pte, val) (!(pte_val(pte) & (val))) pte 223 arch/arm/include/asm/pgtable.h #define pte_none(pte) (!pte_val(pte)) pte 224 arch/arm/include/asm/pgtable.h #define pte_present(pte) (pte_isset((pte), L_PTE_PRESENT)) pte 225 arch/arm/include/asm/pgtable.h #define pte_valid(pte) (pte_isset((pte), L_PTE_VALID)) pte 226 arch/arm/include/asm/pgtable.h #define pte_accessible(mm, pte) (mm_tlb_flush_pending(mm) ? pte_present(pte) : pte_valid(pte)) pte 227 arch/arm/include/asm/pgtable.h #define pte_write(pte) (pte_isclear((pte), L_PTE_RDONLY)) pte 228 arch/arm/include/asm/pgtable.h #define pte_dirty(pte) (pte_isset((pte), L_PTE_DIRTY)) pte 229 arch/arm/include/asm/pgtable.h #define pte_young(pte) (pte_isset((pte), L_PTE_YOUNG)) pte 230 arch/arm/include/asm/pgtable.h #define pte_exec(pte) (pte_isclear((pte), L_PTE_XN)) pte 232 arch/arm/include/asm/pgtable.h #define pte_valid_user(pte) \ pte 233 arch/arm/include/asm/pgtable.h (pte_valid(pte) && pte_isset((pte), L_PTE_USER) && pte_young(pte)) pte 235 arch/arm/include/asm/pgtable.h static inline bool pte_access_permitted(pte_t pte, bool write) pte 243 arch/arm/include/asm/pgtable.h return (pte_val(pte) & mask) == needed; pte 269 arch/arm/include/asm/pgtable.h static inline pte_t clear_pte_bit(pte_t pte, pgprot_t prot) pte 271 arch/arm/include/asm/pgtable.h pte_val(pte) &= ~pgprot_val(prot); pte 272 arch/arm/include/asm/pgtable.h return pte; pte 275 arch/arm/include/asm/pgtable.h static inline pte_t set_pte_bit(pte_t pte, pgprot_t prot) pte 277 arch/arm/include/asm/pgtable.h pte_val(pte) |= pgprot_val(prot); pte 278 arch/arm/include/asm/pgtable.h return pte; pte 281 arch/arm/include/asm/pgtable.h static inline pte_t pte_wrprotect(pte_t pte) pte 283 arch/arm/include/asm/pgtable.h return set_pte_bit(pte, __pgprot(L_PTE_RDONLY)); pte 286 arch/arm/include/asm/pgtable.h static inline pte_t pte_mkwrite(pte_t pte) pte 288 arch/arm/include/asm/pgtable.h return clear_pte_bit(pte, __pgprot(L_PTE_RDONLY)); pte 291 arch/arm/include/asm/pgtable.h static inline pte_t pte_mkclean(pte_t pte) pte 293 arch/arm/include/asm/pgtable.h return clear_pte_bit(pte, __pgprot(L_PTE_DIRTY)); pte 296 arch/arm/include/asm/pgtable.h static inline pte_t pte_mkdirty(pte_t pte) pte 298 arch/arm/include/asm/pgtable.h return set_pte_bit(pte, __pgprot(L_PTE_DIRTY)); pte 301 arch/arm/include/asm/pgtable.h static inline pte_t pte_mkold(pte_t pte) pte 303 arch/arm/include/asm/pgtable.h return clear_pte_bit(pte, __pgprot(L_PTE_YOUNG)); pte 306 arch/arm/include/asm/pgtable.h static inline pte_t pte_mkyoung(pte_t pte) pte 308 arch/arm/include/asm/pgtable.h return set_pte_bit(pte, __pgprot(L_PTE_YOUNG)); pte 311 arch/arm/include/asm/pgtable.h static inline pte_t pte_mkexec(pte_t pte) pte 313 arch/arm/include/asm/pgtable.h return clear_pte_bit(pte, __pgprot(L_PTE_XN)); pte 316 arch/arm/include/asm/pgtable.h static inline pte_t pte_mknexec(pte_t pte) pte 318 arch/arm/include/asm/pgtable.h return set_pte_bit(pte, __pgprot(L_PTE_XN)); pte 321 arch/arm/include/asm/pgtable.h static inline pte_t pte_modify(pte_t pte, pgprot_t newprot) pte 325 arch/arm/include/asm/pgtable.h pte_val(pte) = (pte_val(pte) & ~mask) | (pgprot_val(newprot) & mask); pte 326 arch/arm/include/asm/pgtable.h return pte; pte 349 arch/arm/include/asm/pgtable.h #define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val(pte) }) pte 70 arch/arm/include/asm/proc-fns.h void (*set_pte_ext)(pte_t *ptep, pte_t pte); pte 72 arch/arm/include/asm/proc-fns.h void (*set_pte_ext)(pte_t *ptep, pte_t pte, unsigned int ext); pte 92 arch/arm/include/asm/proc-fns.h extern void cpu_set_pte_ext(pte_t *ptep, pte_t pte); pte 94 arch/arm/include/asm/proc-fns.h extern void cpu_set_pte_ext(pte_t *ptep, pte_t pte, unsigned int ext); pte 45 arch/arm/include/asm/tlb.h __pte_free_tlb(struct mmu_gather *tlb, pgtable_t pte, unsigned long addr) pte 47 arch/arm/include/asm/tlb.h pgtable_pte_page_dtor(pte); pte 58 arch/arm/include/asm/tlb.h tlb_remove_table(tlb, pte); pte 14 arch/arm/kernel/efi.c pte_t pte = *ptep; pte 17 arch/arm/kernel/efi.c pte = set_pte_bit(pte, __pgprot(L_PTE_RDONLY)); pte 19 arch/arm/kernel/efi.c pte = set_pte_bit(pte, __pgprot(L_PTE_XN)); pte 20 arch/arm/kernel/efi.c set_pte_ext(ptep, pte, PTE_EXT_NG); pte 744 arch/arm/kernel/traps.c void __pte_error(const char *file, int line, pte_t pte) pte 746 arch/arm/kernel/traps.c pr_err("%s:%d: bad pte %08llx.\n", file, line, (long long)pte_val(pte)); pte 28 arch/arm/lib/uaccess_with_memcpy.c pte_t *pte; pte 71 arch/arm/lib/uaccess_with_memcpy.c pte = pte_offset_map_lock(current->mm, pmd, addr, &ptl); pte 72 arch/arm/lib/uaccess_with_memcpy.c if (unlikely(!pte_present(*pte) || !pte_young(*pte) || pte 73 arch/arm/lib/uaccess_with_memcpy.c !pte_write(*pte) || !pte_dirty(*pte))) { pte 74 arch/arm/lib/uaccess_with_memcpy.c pte_unmap_unlock(pte, ptl); pte 78 arch/arm/lib/uaccess_with_memcpy.c *ptep = pte; pte 101 arch/arm/lib/uaccess_with_memcpy.c pte_t *pte; pte 105 arch/arm/lib/uaccess_with_memcpy.c while (!pin_page_for_write(to, &pte, &ptl)) { pte 125 arch/arm/lib/uaccess_with_memcpy.c if (pte) pte 126 arch/arm/lib/uaccess_with_memcpy.c pte_unmap_unlock(pte, ptl); pte 170 arch/arm/lib/uaccess_with_memcpy.c pte_t *pte; pte 174 arch/arm/lib/uaccess_with_memcpy.c while (!pin_page_for_write(addr, &pte, &ptl)) { pte 191 arch/arm/lib/uaccess_with_memcpy.c if (pte) pte 192 arch/arm/lib/uaccess_with_memcpy.c pte_unmap_unlock(pte, ptl); pte 464 arch/arm/mm/dma-mapping.c static int __dma_update_pte(pte_t *pte, unsigned long addr, void *data) pte 469 arch/arm/mm/dma-mapping.c set_pte_ext(pte, mk_pte(page, prot), 0); pte 305 arch/arm/mm/dump.c pte_t *pte = pte_offset_kernel(pmd, 0); pte 309 arch/arm/mm/dump.c for (i = 0; i < PTRS_PER_PTE; i++, pte++) { pte 311 arch/arm/mm/dump.c note_page(st, addr, 4, pte_val(*pte), domain); pte 96 arch/arm/mm/fault-armv.c pte_t *pte; pte 117 arch/arm/mm/fault-armv.c pte = pte_offset_map(pmd, address); pte 120 arch/arm/mm/fault-armv.c ret = do_adjust_pte(vma, address, pfn, pte); pte 123 arch/arm/mm/fault-armv.c pte_unmap(pte); pte 48 arch/arm/mm/fault.c pte_t *pte; pte 86 arch/arm/mm/fault.c pte = pte_offset_map(pmd, addr); pte 87 arch/arm/mm/fault.c pr_cont(", *pte=%08llx", (long long)pte_val(*pte)); pte 90 arch/arm/mm/fault.c (long long)pte_val(pte[PTE_HWTABLE_PTRS])); pte 92 arch/arm/mm/fault.c pte_unmap(pte); pte 18 arch/arm/mm/highmem.c static inline void set_fixmap_pte(int idx, pte_t pte) pte 23 arch/arm/mm/highmem.c set_pte_ext(ptep, pte, 0); pte 26 arch/arm/mm/mm.h static inline void set_top_pte(unsigned long va, pte_t pte) pte 29 arch/arm/mm/mm.h set_pte_ext(ptep, pte, 0); pte 77 arch/arm/mm/mmu.c pteval_t pte; pte 94 arch/arm/mm/mmu.c .pte = L_PTE_MT_UNCACHED, pte 100 arch/arm/mm/mmu.c .pte = L_PTE_MT_BUFFERABLE, pte 106 arch/arm/mm/mmu.c .pte = L_PTE_MT_WRITETHROUGH, pte 112 arch/arm/mm/mmu.c .pte = L_PTE_MT_WRITEBACK, pte 118 arch/arm/mm/mmu.c .pte = L_PTE_MT_WRITEALLOC, pte 409 arch/arm/mm/mmu.c pte_t *pte = pte_offset_fixmap(pmd_off_k(vaddr), vaddr); pte 422 arch/arm/mm/mmu.c set_pte_at(NULL, vaddr, pte, pte 425 arch/arm/mm/mmu.c pte_clear(NULL, vaddr, pte); pte 560 arch/arm/mm/mmu.c vecs_pgprot = kern_pgprot = user_pgprot = cp->pte; pte 744 arch/arm/mm/mmu.c pte_t *pte = alloc(PTE_HWTABLE_OFF + PTE_HWTABLE_SIZE); pte 745 arch/arm/mm/mmu.c __pmd_populate(pmd, __pa(pte), prot); pte 763 arch/arm/mm/mmu.c pte_t *pte = arm_pte_alloc(pmd, addr, type->prot_l1, alloc); pte 765 arch/arm/mm/mmu.c set_pte_ext(pte, pfn_pte(pfn, __pgprot(type->prot_pte)), pte 768 arch/arm/mm/mmu.c } while (pte++, addr += PAGE_SIZE, addr != end); pte 1620 arch/arm/mm/mmu.c pte_t *pte; pte 1624 arch/arm/mm/mmu.c pte = pte_offset_early_fixmap(pmd_off_k(map.virtual), map.virtual); pte 1627 arch/arm/mm/mmu.c if (pte_none(*pte) || pte 1628 arch/arm/mm/mmu.c (pte_val(*pte) & L_PTE_MT_MASK) != L_PTE_MT_DEV_SHARED) pte 1631 arch/arm/mm/mmu.c map.pfn = pte_pfn(*pte); pte 20 arch/arm/mm/pageattr.c pte_t pte = *ptep; pte 22 arch/arm/mm/pageattr.c pte = clear_pte_bit(pte, cdata->clear_mask); pte 23 arch/arm/mm/pageattr.c pte = set_pte_bit(pte, cdata->set_mask); pte 25 arch/arm/mm/pageattr.c set_pte_ext(ptep, pte, 0); pte 121 arch/arm/mm/pgd.c pgtable_t pte; pte 138 arch/arm/mm/pgd.c pte = pmd_pgtable(*pmd); pte 140 arch/arm/mm/pgd.c pte_free(mm, pte); pte 574 arch/arm64/include/asm/assembler.h .macro phys_to_pte, pte, phys pte 587 arch/arm64/include/asm/assembler.h .macro pte_to_phys, phys, pte pte 42 arch/arm64/include/asm/hugetlb.h pte_t *ptep, pte_t pte); pte 46 arch/arm64/include/asm/hugetlb.h pte_t pte, int dirty); pte 60 arch/arm64/include/asm/hugetlb.h pte_t *ptep, pte_t pte, unsigned long sz); pte 431 arch/arm64/include/asm/kvm_host.h int kvm_set_spte_hva(struct kvm *kvm, unsigned long hva, pte_t pte); pte 188 arch/arm64/include/asm/kvm_mmu.h static inline pte_t kvm_s2pte_mkwrite(pte_t pte) pte 190 arch/arm64/include/asm/kvm_mmu.h pte_val(pte) |= PTE_S2_RDWR; pte 191 arch/arm64/include/asm/kvm_mmu.h return pte; pte 206 arch/arm64/include/asm/kvm_mmu.h static inline pte_t kvm_s2pte_mkexec(pte_t pte) pte 208 arch/arm64/include/asm/kvm_mmu.h pte_val(pte) &= ~PTE_S2_XN; pte 209 arch/arm64/include/asm/kvm_mmu.h return pte; pte 341 arch/arm64/include/asm/kvm_mmu.h static inline void __kvm_flush_dcache_pte(pte_t pte) pte 344 arch/arm64/include/asm/kvm_mmu.h struct page *page = pte_page(pte); pte 22 arch/arm64/include/asm/pgtable-types.h typedef struct { pteval_t pte; } pte_t; pte 23 arch/arm64/include/asm/pgtable-types.h #define pte_val(x) ((x).pte) pte 50 arch/arm64/include/asm/pgtable.h #define pte_ERROR(pte) __pte_error(__FILE__, __LINE__, pte_val(pte)) pte 57 arch/arm64/include/asm/pgtable.h #define __pte_to_phys(pte) \ pte 58 arch/arm64/include/asm/pgtable.h ((pte_val(pte) & PTE_ADDR_LOW) | ((pte_val(pte) & PTE_ADDR_HIGH) << 36)) pte 61 arch/arm64/include/asm/pgtable.h #define __pte_to_phys(pte) (pte_val(pte) & PTE_ADDR_MASK) pte 65 arch/arm64/include/asm/pgtable.h #define pte_pfn(pte) (__pte_to_phys(pte) >> PAGE_SHIFT) pte 69 arch/arm64/include/asm/pgtable.h #define pte_none(pte) (!pte_val(pte)) pte 71 arch/arm64/include/asm/pgtable.h #define pte_page(pte) (pfn_to_page(pte_pfn(pte))) pte 76 arch/arm64/include/asm/pgtable.h #define pte_present(pte) (!!(pte_val(pte) & (PTE_VALID | PTE_PROT_NONE))) pte 77 arch/arm64/include/asm/pgtable.h #define pte_young(pte) (!!(pte_val(pte) & PTE_AF)) pte 78 arch/arm64/include/asm/pgtable.h #define pte_special(pte) (!!(pte_val(pte) & PTE_SPECIAL)) pte 79 arch/arm64/include/asm/pgtable.h #define pte_write(pte) (!!(pte_val(pte) & PTE_WRITE)) pte 80 arch/arm64/include/asm/pgtable.h #define pte_user_exec(pte) (!(pte_val(pte) & PTE_UXN)) pte 81 arch/arm64/include/asm/pgtable.h #define pte_cont(pte) (!!(pte_val(pte) & PTE_CONT)) pte 82 arch/arm64/include/asm/pgtable.h #define pte_devmap(pte) (!!(pte_val(pte) & PTE_DEVMAP)) pte 94 arch/arm64/include/asm/pgtable.h #define pte_hw_dirty(pte) (pte_write(pte) && !(pte_val(pte) & PTE_RDONLY)) pte 95 arch/arm64/include/asm/pgtable.h #define pte_sw_dirty(pte) (!!(pte_val(pte) & PTE_DIRTY)) pte 96 arch/arm64/include/asm/pgtable.h #define pte_dirty(pte) (pte_sw_dirty(pte) || pte_hw_dirty(pte)) pte 98 arch/arm64/include/asm/pgtable.h #define pte_valid(pte) (!!(pte_val(pte) & PTE_VALID)) pte 99 arch/arm64/include/asm/pgtable.h #define pte_valid_not_user(pte) \ pte 100 arch/arm64/include/asm/pgtable.h ((pte_val(pte) & (PTE_VALID | PTE_USER)) == PTE_VALID) pte 101 arch/arm64/include/asm/pgtable.h #define pte_valid_young(pte) \ pte 102 arch/arm64/include/asm/pgtable.h ((pte_val(pte) & (PTE_VALID | PTE_AF)) == (PTE_VALID | PTE_AF)) pte 103 arch/arm64/include/asm/pgtable.h #define pte_valid_user(pte) \ pte 104 arch/arm64/include/asm/pgtable.h ((pte_val(pte) & (PTE_VALID | PTE_USER)) == (PTE_VALID | PTE_USER)) pte 111 arch/arm64/include/asm/pgtable.h #define pte_accessible(mm, pte) \ pte 112 arch/arm64/include/asm/pgtable.h (mm_tlb_flush_pending(mm) ? pte_present(pte) : pte_valid_young(pte)) pte 119 arch/arm64/include/asm/pgtable.h #define pte_access_permitted(pte, write) \ pte 120 arch/arm64/include/asm/pgtable.h (pte_valid_user(pte) && (!(write) || pte_write(pte))) pte 126 arch/arm64/include/asm/pgtable.h static inline pte_t clear_pte_bit(pte_t pte, pgprot_t prot) pte 128 arch/arm64/include/asm/pgtable.h pte_val(pte) &= ~pgprot_val(prot); pte 129 arch/arm64/include/asm/pgtable.h return pte; pte 132 arch/arm64/include/asm/pgtable.h static inline pte_t set_pte_bit(pte_t pte, pgprot_t prot) pte 134 arch/arm64/include/asm/pgtable.h pte_val(pte) |= pgprot_val(prot); pte 135 arch/arm64/include/asm/pgtable.h return pte; pte 138 arch/arm64/include/asm/pgtable.h static inline pte_t pte_wrprotect(pte_t pte) pte 140 arch/arm64/include/asm/pgtable.h pte = clear_pte_bit(pte, __pgprot(PTE_WRITE)); pte 141 arch/arm64/include/asm/pgtable.h pte = set_pte_bit(pte, __pgprot(PTE_RDONLY)); pte 142 arch/arm64/include/asm/pgtable.h return pte; pte 145 arch/arm64/include/asm/pgtable.h static inline pte_t pte_mkwrite(pte_t pte) pte 147 arch/arm64/include/asm/pgtable.h pte = set_pte_bit(pte, __pgprot(PTE_WRITE)); pte 148 arch/arm64/include/asm/pgtable.h pte = clear_pte_bit(pte, __pgprot(PTE_RDONLY)); pte 149 arch/arm64/include/asm/pgtable.h return pte; pte 152 arch/arm64/include/asm/pgtable.h static inline pte_t pte_mkclean(pte_t pte) pte 154 arch/arm64/include/asm/pgtable.h pte = clear_pte_bit(pte, __pgprot(PTE_DIRTY)); pte 155 arch/arm64/include/asm/pgtable.h pte = set_pte_bit(pte, __pgprot(PTE_RDONLY)); pte 157 arch/arm64/include/asm/pgtable.h return pte; pte 160 arch/arm64/include/asm/pgtable.h static inline pte_t pte_mkdirty(pte_t pte) pte 162 arch/arm64/include/asm/pgtable.h pte = set_pte_bit(pte, __pgprot(PTE_DIRTY)); pte 164 arch/arm64/include/asm/pgtable.h if (pte_write(pte)) pte 165 arch/arm64/include/asm/pgtable.h pte = clear_pte_bit(pte, __pgprot(PTE_RDONLY)); pte 167 arch/arm64/include/asm/pgtable.h return pte; pte 170 arch/arm64/include/asm/pgtable.h static inline pte_t pte_mkold(pte_t pte) pte 172 arch/arm64/include/asm/pgtable.h return clear_pte_bit(pte, __pgprot(PTE_AF)); pte 175 arch/arm64/include/asm/pgtable.h static inline pte_t pte_mkyoung(pte_t pte) pte 177 arch/arm64/include/asm/pgtable.h return set_pte_bit(pte, __pgprot(PTE_AF)); pte 180 arch/arm64/include/asm/pgtable.h static inline pte_t pte_mkspecial(pte_t pte) pte 182 arch/arm64/include/asm/pgtable.h return set_pte_bit(pte, __pgprot(PTE_SPECIAL)); pte 185 arch/arm64/include/asm/pgtable.h static inline pte_t pte_mkcont(pte_t pte) pte 187 arch/arm64/include/asm/pgtable.h pte = set_pte_bit(pte, __pgprot(PTE_CONT)); pte 188 arch/arm64/include/asm/pgtable.h return set_pte_bit(pte, __pgprot(PTE_TYPE_PAGE)); pte 191 arch/arm64/include/asm/pgtable.h static inline pte_t pte_mknoncont(pte_t pte) pte 193 arch/arm64/include/asm/pgtable.h return clear_pte_bit(pte, __pgprot(PTE_CONT)); pte 196 arch/arm64/include/asm/pgtable.h static inline pte_t pte_mkpresent(pte_t pte) pte 198 arch/arm64/include/asm/pgtable.h return set_pte_bit(pte, __pgprot(PTE_VALID)); pte 206 arch/arm64/include/asm/pgtable.h static inline pte_t pte_mkdevmap(pte_t pte) pte 208 arch/arm64/include/asm/pgtable.h return set_pte_bit(pte, __pgprot(PTE_DEVMAP | PTE_SPECIAL)); pte 211 arch/arm64/include/asm/pgtable.h static inline void set_pte(pte_t *ptep, pte_t pte) pte 213 arch/arm64/include/asm/pgtable.h WRITE_ONCE(*ptep, pte); pte 219 arch/arm64/include/asm/pgtable.h if (pte_valid_not_user(pte)) { pte 244 arch/arm64/include/asm/pgtable.h pte_t pte) pte 253 arch/arm64/include/asm/pgtable.h if (!pte_valid(old_pte) || !pte_valid(pte)) pte 263 arch/arm64/include/asm/pgtable.h VM_WARN_ONCE(!pte_young(pte), pte 265 arch/arm64/include/asm/pgtable.h __func__, pte_val(old_pte), pte_val(pte)); pte 266 arch/arm64/include/asm/pgtable.h VM_WARN_ONCE(pte_write(old_pte) && !pte_dirty(pte), pte 268 arch/arm64/include/asm/pgtable.h __func__, pte_val(old_pte), pte_val(pte)); pte 272 arch/arm64/include/asm/pgtable.h pte_t *ptep, pte_t pte) pte 274 arch/arm64/include/asm/pgtable.h if (pte_present(pte) && pte_user_exec(pte) && !pte_special(pte)) pte 275 arch/arm64/include/asm/pgtable.h __sync_icache_dcache(pte); pte 277 arch/arm64/include/asm/pgtable.h __check_racy_pte_update(mm, ptep, pte); pte 279 arch/arm64/include/asm/pgtable.h set_pte(ptep, pte); pte 285 arch/arm64/include/asm/pgtable.h #define pte_mkhuge(pte) (__pte(pte_val(pte) & ~PTE_TABLE_BIT)) pte 306 arch/arm64/include/asm/pgtable.h static inline pud_t pte_pud(pte_t pte) pte 308 arch/arm64/include/asm/pgtable.h return __pud(pte_val(pte)); pte 321 arch/arm64/include/asm/pgtable.h static inline pmd_t pte_pmd(pte_t pte) pte 323 arch/arm64/include/asm/pgtable.h return __pmd(pte_val(pte)); pte 340 arch/arm64/include/asm/pgtable.h static inline int pte_protnone(pte_t pte) pte 342 arch/arm64/include/asm/pgtable.h return (pte_val(pte) & (PTE_VALID | PTE_PROT_NONE)) == PTE_PROT_NONE; pte 496 arch/arm64/include/asm/pgtable.h static inline void pte_unmap(pte_t *pte) { } pte 658 arch/arm64/include/asm/pgtable.h static inline pte_t pte_modify(pte_t pte, pgprot_t newprot) pte 663 arch/arm64/include/asm/pgtable.h if (pte_hw_dirty(pte)) pte 664 arch/arm64/include/asm/pgtable.h pte = pte_mkdirty(pte); pte 665 arch/arm64/include/asm/pgtable.h pte_val(pte) = (pte_val(pte) & ~mask) | (pgprot_val(newprot) & mask); pte 666 arch/arm64/include/asm/pgtable.h return pte; pte 705 arch/arm64/include/asm/pgtable.h pte_t old_pte, pte; pte 707 arch/arm64/include/asm/pgtable.h pte = READ_ONCE(*ptep); pte 709 arch/arm64/include/asm/pgtable.h old_pte = pte; pte 710 arch/arm64/include/asm/pgtable.h pte = pte_mkold(pte); pte 711 arch/arm64/include/asm/pgtable.h pte_val(pte) = cmpxchg_relaxed(&pte_val(*ptep), pte 712 arch/arm64/include/asm/pgtable.h pte_val(old_pte), pte_val(pte)); pte 713 arch/arm64/include/asm/pgtable.h } while (pte_val(pte) != pte_val(old_pte)); pte 715 arch/arm64/include/asm/pgtable.h return pte_young(pte); pte 779 arch/arm64/include/asm/pgtable.h pte_t old_pte, pte; pte 781 arch/arm64/include/asm/pgtable.h pte = READ_ONCE(*ptep); pte 783 arch/arm64/include/asm/pgtable.h old_pte = pte; pte 788 arch/arm64/include/asm/pgtable.h if (pte_hw_dirty(pte)) pte 789 arch/arm64/include/asm/pgtable.h pte = pte_mkdirty(pte); pte 790 arch/arm64/include/asm/pgtable.h pte = pte_wrprotect(pte); pte 791 arch/arm64/include/asm/pgtable.h pte_val(pte) = cmpxchg_relaxed(&pte_val(*ptep), pte 792 arch/arm64/include/asm/pgtable.h pte_val(old_pte), pte_val(pte)); pte 793 arch/arm64/include/asm/pgtable.h } while (pte_val(pte) != pte_val(old_pte)); pte 830 arch/arm64/include/asm/pgtable.h #define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val(pte) }) pte 44 arch/arm64/include/asm/tlb.h static inline void __pte_free_tlb(struct mmu_gather *tlb, pgtable_t pte, pte 47 arch/arm64/include/asm/tlb.h pgtable_pte_page_dtor(pte); pte 48 arch/arm64/include/asm/tlb.h tlb_remove_table(tlb, pte); pte 88 arch/arm64/kernel/efi.c pte_t pte = READ_ONCE(*ptep); pte 91 arch/arm64/kernel/efi.c pte = set_pte_bit(pte, __pgprot(PTE_RDONLY)); pte 93 arch/arm64/kernel/efi.c pte = set_pte_bit(pte, __pgprot(PTE_PXN)); pte 94 arch/arm64/kernel/efi.c set_pte(ptep, pte); pte 344 arch/arm64/kernel/hibernate.c pte_t pte = READ_ONCE(*src_ptep); pte 346 arch/arm64/kernel/hibernate.c if (pte_valid(pte)) { pte 352 arch/arm64/kernel/hibernate.c set_pte(dst_ptep, pte_mkwrite(pte)); pte 353 arch/arm64/kernel/hibernate.c } else if (debug_pagealloc_enabled() && !pte_none(pte)) { pte 363 arch/arm64/kernel/hibernate.c BUG_ON(!pfn_valid(pte_pfn(pte))); pte 365 arch/arm64/kernel/hibernate.c set_pte(dst_ptep, pte_mkpresent(pte_mkwrite(pte))); pte 161 arch/arm64/mm/fault.c pte_t *ptep, pte; pte 179 arch/arm64/mm/fault.c pte = READ_ONCE(*ptep); pte 180 arch/arm64/mm/fault.c pr_cont(", pte=%016llx", pte_val(pte)); pte 202 arch/arm64/mm/fault.c pte_t pte = READ_ONCE(*ptep); pte 204 arch/arm64/mm/fault.c if (pte_same(pte, entry)) pte 217 arch/arm64/mm/fault.c pteval = pte_val(pte); pte 54 arch/arm64/mm/flush.c void __sync_icache_dcache(pte_t pte) pte 56 arch/arm64/mm/flush.c struct page *page = pte_page(pte); pte 59 arch/arm64/mm/hugetlbpage.c static inline pgprot_t pte_pgprot(pte_t pte) pte 61 arch/arm64/mm/hugetlbpage.c unsigned long pfn = pte_pfn(pte); pte 63 arch/arm64/mm/hugetlbpage.c return __pgprot(pte_val(pfn_pte(pfn, __pgprot(0))) ^ pte_val(pte)); pte 128 arch/arm64/mm/hugetlbpage.c pte_t pte = ptep_get_and_clear(mm, addr, ptep); pte 135 arch/arm64/mm/hugetlbpage.c if (pte_dirty(pte)) pte 138 arch/arm64/mm/hugetlbpage.c if (pte_young(pte)) pte 174 arch/arm64/mm/hugetlbpage.c pte_t *ptep, pte_t pte) pte 186 arch/arm64/mm/hugetlbpage.c WARN_ON(!pte_present(pte)); pte 188 arch/arm64/mm/hugetlbpage.c if (!pte_cont(pte)) { pte 189 arch/arm64/mm/hugetlbpage.c set_pte_at(mm, addr, ptep, pte); pte 194 arch/arm64/mm/hugetlbpage.c pfn = pte_pfn(pte); pte 196 arch/arm64/mm/hugetlbpage.c hugeprot = pte_pgprot(pte); pte 205 arch/arm64/mm/hugetlbpage.c pte_t *ptep, pte_t pte, unsigned long sz) pte 213 arch/arm64/mm/hugetlbpage.c set_pte(ptep, pte); pte 349 arch/arm64/mm/hugetlbpage.c static int __cont_access_flags_changed(pte_t *ptep, pte_t pte, int ncontig) pte 353 arch/arm64/mm/hugetlbpage.c if (pte_write(pte) != pte_write(huge_ptep_get(ptep))) pte 359 arch/arm64/mm/hugetlbpage.c if (pte_dirty(pte) != pte_dirty(orig_pte)) pte 362 arch/arm64/mm/hugetlbpage.c if (pte_young(pte) != pte_young(orig_pte)) pte 371 arch/arm64/mm/hugetlbpage.c pte_t pte, int dirty) pte 375 arch/arm64/mm/hugetlbpage.c unsigned long pfn = pte_pfn(pte), dpfn; pte 379 arch/arm64/mm/hugetlbpage.c if (!pte_cont(pte)) pte 380 arch/arm64/mm/hugetlbpage.c return ptep_set_access_flags(vma, addr, ptep, pte, dirty); pte 385 arch/arm64/mm/hugetlbpage.c if (!__cont_access_flags_changed(ptep, pte, ncontig)) pte 392 arch/arm64/mm/hugetlbpage.c pte = pte_mkdirty(pte); pte 395 arch/arm64/mm/hugetlbpage.c pte = pte_mkyoung(pte); pte 397 arch/arm64/mm/hugetlbpage.c hugeprot = pte_pgprot(pte); pte 411 arch/arm64/mm/hugetlbpage.c pte_t pte; pte 421 arch/arm64/mm/hugetlbpage.c pte = get_clear_flush(mm, addr, ptep, pgsize, ncontig); pte 422 arch/arm64/mm/hugetlbpage.c pte = pte_wrprotect(pte); pte 424 arch/arm64/mm/hugetlbpage.c hugeprot = pte_pgprot(pte); pte 425 arch/arm64/mm/hugetlbpage.c pfn = pte_pfn(pte); pte 696 arch/arm64/mm/mmu.c pte_t *ptep, pte; pte 722 arch/arm64/mm/mmu.c pte = READ_ONCE(*ptep); pte 723 arch/arm64/mm/mmu.c if (pte_none(pte)) pte 726 arch/arm64/mm/mmu.c return pfn_valid(pte_pfn(pte)); pte 25 arch/arm64/mm/pageattr.c pte_t pte = READ_ONCE(*ptep); pte 27 arch/arm64/mm/pageattr.c pte = clear_pte_bit(pte, cdata->clear_mask); pte 28 arch/arm64/mm/pageattr.c pte = set_pte_bit(pte, cdata->set_mask); pte 30 arch/arm64/mm/pageattr.c set_pte(ptep, pte); pte 47 arch/c6x/include/asm/pgtable.h #define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val(pte) }) pte 33 arch/csky/abiv2/cacheflush.c pte_t *pte) pte 38 arch/csky/abiv2/cacheflush.c pfn = pte_pfn(*pte); pte 15 arch/csky/include/asm/pgalloc.h pte_t *pte) pte 17 arch/csky/include/asm/pgalloc.h set_pmd(pmd, __pmd(__pa(pte))); pte 21 arch/csky/include/asm/pgalloc.h pgtable_t pte) pte 23 arch/csky/include/asm/pgalloc.h set_pmd(pmd, __pmd(__pa(page_address(pte)))); pte 32 arch/csky/include/asm/pgalloc.h pte_t *pte; pte 35 arch/csky/include/asm/pgalloc.h pte = (pte_t *) __get_free_page(GFP_KERNEL); pte 36 arch/csky/include/asm/pgalloc.h if (!pte) pte 40 arch/csky/include/asm/pgalloc.h (pte + i)->pte_low = _PAGE_GLOBAL; pte 42 arch/csky/include/asm/pgalloc.h return pte; pte 72 arch/csky/include/asm/pgalloc.h #define __pte_free_tlb(tlb, pte, address) \ pte 74 arch/csky/include/asm/pgalloc.h pgtable_pte_page_dtor(pte); \ pte 75 arch/csky/include/asm/pgalloc.h tlb_remove_page(tlb, pte); \ pte 49 arch/csky/include/asm/pgtable.h #define pte_none(pte) (!(pte_val(pte) & ~_PAGE_GLOBAL)) pte 50 arch/csky/include/asm/pgtable.h #define pte_present(pte) (pte_val(pte) & _PAGE_PRESENT) pte 61 arch/csky/include/asm/pgtable.h #define pte_unmap(pte) ((void)(pte)) pte 67 arch/csky/include/asm/pgtable.h #define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val(pte) }) pte 113 arch/csky/include/asm/pgtable.h static inline int pte_special(pte_t pte) { return 0; } pte 114 arch/csky/include/asm/pgtable.h static inline pte_t pte_mkspecial(pte_t pte) { return pte; } pte 116 arch/csky/include/asm/pgtable.h static inline void set_pte(pte_t *p, pte_t pte) pte 118 arch/csky/include/asm/pgtable.h *p = pte; pte 173 arch/csky/include/asm/pgtable.h static inline int pte_read(pte_t pte) pte 175 arch/csky/include/asm/pgtable.h return pte.pte_low & _PAGE_READ; pte 178 arch/csky/include/asm/pgtable.h static inline int pte_write(pte_t pte) pte 180 arch/csky/include/asm/pgtable.h return (pte).pte_low & _PAGE_WRITE; pte 183 arch/csky/include/asm/pgtable.h static inline int pte_dirty(pte_t pte) pte 185 arch/csky/include/asm/pgtable.h return (pte).pte_low & _PAGE_MODIFIED; pte 188 arch/csky/include/asm/pgtable.h static inline int pte_young(pte_t pte) pte 190 arch/csky/include/asm/pgtable.h return (pte).pte_low & _PAGE_ACCESSED; pte 193 arch/csky/include/asm/pgtable.h static inline pte_t pte_wrprotect(pte_t pte) pte 195 arch/csky/include/asm/pgtable.h pte_val(pte) &= ~(_PAGE_WRITE | _PAGE_DIRTY); pte 196 arch/csky/include/asm/pgtable.h return pte; pte 199 arch/csky/include/asm/pgtable.h static inline pte_t pte_mkclean(pte_t pte) pte 201 arch/csky/include/asm/pgtable.h pte_val(pte) &= ~(_PAGE_MODIFIED|_PAGE_DIRTY); pte 202 arch/csky/include/asm/pgtable.h return pte; pte 205 arch/csky/include/asm/pgtable.h static inline pte_t pte_mkold(pte_t pte) pte 207 arch/csky/include/asm/pgtable.h pte_val(pte) &= ~(_PAGE_ACCESSED|_PAGE_VALID); pte 208 arch/csky/include/asm/pgtable.h return pte; pte 211 arch/csky/include/asm/pgtable.h static inline pte_t pte_mkwrite(pte_t pte) pte 213 arch/csky/include/asm/pgtable.h pte_val(pte) |= _PAGE_WRITE; pte 214 arch/csky/include/asm/pgtable.h if (pte_val(pte) & _PAGE_MODIFIED) pte 215 arch/csky/include/asm/pgtable.h pte_val(pte) |= _PAGE_DIRTY; pte 216 arch/csky/include/asm/pgtable.h return pte; pte 219 arch/csky/include/asm/pgtable.h static inline pte_t pte_mkdirty(pte_t pte) pte 221 arch/csky/include/asm/pgtable.h pte_val(pte) |= _PAGE_MODIFIED; pte 222 arch/csky/include/asm/pgtable.h if (pte_val(pte) & _PAGE_WRITE) pte 223 arch/csky/include/asm/pgtable.h pte_val(pte) |= _PAGE_DIRTY; pte 224 arch/csky/include/asm/pgtable.h return pte; pte 227 arch/csky/include/asm/pgtable.h static inline pte_t pte_mkyoung(pte_t pte) pte 229 arch/csky/include/asm/pgtable.h pte_val(pte) |= _PAGE_ACCESSED; pte 230 arch/csky/include/asm/pgtable.h if (pte_val(pte) & _PAGE_READ) pte 231 arch/csky/include/asm/pgtable.h pte_val(pte) |= _PAGE_VALID; pte 232 arch/csky/include/asm/pgtable.h return pte; pte 281 arch/csky/include/asm/pgtable.h static inline pte_t pte_modify(pte_t pte, pgprot_t newprot) pte 283 arch/csky/include/asm/pgtable.h return __pte((pte_val(pte) & _PAGE_CHG_MASK) | pte 304 arch/csky/include/asm/pgtable.h pte_t *pte); pte 110 arch/csky/mm/highmem.c pte_t *pte; pte 116 arch/csky/mm/highmem.c pte = kmap_pte - (idx - FIX_KMAP_BEGIN); pte 117 arch/csky/mm/highmem.c return pte_page(*pte); pte 127 arch/csky/mm/highmem.c pte_t *pte; pte 143 arch/csky/mm/highmem.c pte = (pte_t *) memblock_alloc_low(PAGE_SIZE, PAGE_SIZE); pte 144 arch/csky/mm/highmem.c if (!pte) pte 149 arch/csky/mm/highmem.c set_pmd(pmd, __pmd(__pa(pte))); pte 150 arch/csky/mm/highmem.c BUG_ON(pte != pte_offset_kernel(pmd, 0)); pte 169 arch/csky/mm/highmem.c pte_t *pte; pte 189 arch/csky/mm/highmem.c pte = pte_offset_kernel(pmd, vaddr); pte 190 arch/csky/mm/highmem.c pkmap_page_table = pte; pte 16 arch/h8300/include/asm/pgtable.h #define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val(pte) }) pte 22 arch/h8300/include/asm/pgtable.h static inline int pte_file(pte_t pte) { return 0; } pte 69 arch/hexagon/include/asm/page.h typedef struct { unsigned long pte; } pte_t; pte 74 arch/hexagon/include/asm/page.h #define pte_val(x) ((x).pte) pte 50 arch/hexagon/include/asm/pgalloc.h pgtable_t pte) pte 56 arch/hexagon/include/asm/pgalloc.h set_pmd(pmd, __pmd(((unsigned long)page_to_pfn(pte) << PAGE_SHIFT) | pte 70 arch/hexagon/include/asm/pgalloc.h pte_t *pte) pte 82 arch/hexagon/include/asm/pgalloc.h set_pmd(pmd, __pmd(((unsigned long)__pa(pte)) | HEXAGON_L1_PTE_SIZE)); pte 90 arch/hexagon/include/asm/pgalloc.h set_pmd(ppmd, __pmd(((unsigned long)__pa(pte)) | HEXAGON_L1_PTE_SIZE)); pte 95 arch/hexagon/include/asm/pgalloc.h #define __pte_free_tlb(tlb, pte, addr) \ pte 97 arch/hexagon/include/asm/pgalloc.h pgtable_pte_page_dtor((pte)); \ pte 98 arch/hexagon/include/asm/pgalloc.h tlb_remove_page((tlb), (pte)); \ pte 161 arch/hexagon/include/asm/pgtable.h #define pte_special(pte) 0 pte 162 arch/hexagon/include/asm/pgtable.h #define pte_mkspecial(pte) (pte) pte 166 arch/hexagon/include/asm/pgtable.h #define pte_mkhuge(pte) __pte((pte_val(pte) & ~0x3) | HVM_HUGEPAGE_SIZE) pte 173 arch/hexagon/include/asm/pgtable.h extern void sync_icache_dcache(pte_t pte); pte 175 arch/hexagon/include/asm/pgtable.h #define pte_present_exec_user(pte) \ pte 176 arch/hexagon/include/asm/pgtable.h ((pte_val(pte) & (_PAGE_EXECUTE | _PAGE_USER)) == \ pte 281 arch/hexagon/include/asm/pgtable.h static inline int pte_none(pte_t pte) pte 283 arch/hexagon/include/asm/pgtable.h return pte_val(pte) == _NULL_PTE; pte 289 arch/hexagon/include/asm/pgtable.h static inline int pte_present(pte_t pte) pte 291 arch/hexagon/include/asm/pgtable.h return pte_val(pte) & _PAGE_PRESENT; pte 301 arch/hexagon/include/asm/pgtable.h static inline pte_t pte_mkold(pte_t pte) pte 303 arch/hexagon/include/asm/pgtable.h pte_val(pte) &= ~_PAGE_ACCESSED; pte 304 arch/hexagon/include/asm/pgtable.h return pte; pte 308 arch/hexagon/include/asm/pgtable.h static inline pte_t pte_mkyoung(pte_t pte) pte 310 arch/hexagon/include/asm/pgtable.h pte_val(pte) |= _PAGE_ACCESSED; pte 311 arch/hexagon/include/asm/pgtable.h return pte; pte 315 arch/hexagon/include/asm/pgtable.h static inline pte_t pte_mkclean(pte_t pte) pte 317 arch/hexagon/include/asm/pgtable.h pte_val(pte) &= ~_PAGE_DIRTY; pte 318 arch/hexagon/include/asm/pgtable.h return pte; pte 322 arch/hexagon/include/asm/pgtable.h static inline pte_t pte_mkdirty(pte_t pte) pte 324 arch/hexagon/include/asm/pgtable.h pte_val(pte) |= _PAGE_DIRTY; pte 325 arch/hexagon/include/asm/pgtable.h return pte; pte 329 arch/hexagon/include/asm/pgtable.h static inline int pte_young(pte_t pte) pte 331 arch/hexagon/include/asm/pgtable.h return pte_val(pte) & _PAGE_ACCESSED; pte 335 arch/hexagon/include/asm/pgtable.h static inline int pte_dirty(pte_t pte) pte 337 arch/hexagon/include/asm/pgtable.h return pte_val(pte) & _PAGE_DIRTY; pte 341 arch/hexagon/include/asm/pgtable.h static inline pte_t pte_modify(pte_t pte, pgprot_t prot) pte 343 arch/hexagon/include/asm/pgtable.h pte_val(pte) &= PAGE_MASK; pte 344 arch/hexagon/include/asm/pgtable.h pte_val(pte) |= pgprot_val(prot); pte 345 arch/hexagon/include/asm/pgtable.h return pte; pte 349 arch/hexagon/include/asm/pgtable.h static inline pte_t pte_wrprotect(pte_t pte) pte 351 arch/hexagon/include/asm/pgtable.h pte_val(pte) &= ~_PAGE_WRITE; pte 352 arch/hexagon/include/asm/pgtable.h return pte; pte 356 arch/hexagon/include/asm/pgtable.h static inline pte_t pte_mkwrite(pte_t pte) pte 358 arch/hexagon/include/asm/pgtable.h pte_val(pte) |= _PAGE_WRITE; pte 359 arch/hexagon/include/asm/pgtable.h return pte; pte 363 arch/hexagon/include/asm/pgtable.h static inline pte_t pte_mkexec(pte_t pte) pte 365 arch/hexagon/include/asm/pgtable.h pte_val(pte) |= _PAGE_EXECUTE; pte 366 arch/hexagon/include/asm/pgtable.h return pte; pte 370 arch/hexagon/include/asm/pgtable.h static inline int pte_read(pte_t pte) pte 372 arch/hexagon/include/asm/pgtable.h return pte_val(pte) & _PAGE_READ; pte 376 arch/hexagon/include/asm/pgtable.h static inline int pte_write(pte_t pte) pte 378 arch/hexagon/include/asm/pgtable.h return pte_val(pte) & _PAGE_WRITE; pte 383 arch/hexagon/include/asm/pgtable.h static inline int pte_exec(pte_t pte) pte 385 arch/hexagon/include/asm/pgtable.h return pte_val(pte) & _PAGE_EXECUTE; pte 389 arch/hexagon/include/asm/pgtable.h #define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val(pte) }) pte 398 arch/hexagon/include/asm/pgtable.h #define pte_pfn(pte) (pte_val(pte) >> PAGE_SHIFT) pte 407 arch/hexagon/include/asm/pgtable.h #define set_pte_at(mm, addr, ptep, pte) set_pte(ptep, pte) pte 412 arch/hexagon/include/asm/pgtable.h #define pte_unmap(pte) do { } while (0) pte 413 arch/hexagon/include/asm/pgtable.h #define pte_unmap_nested(pte) do { } while (0) pte 47 arch/hexagon/include/asm/vm_mmu.h #define __HVM_PTE_CVAL(pte) (((pte) & __HVM_PTE_C) >> 6) pte 74 arch/hexagon/mm/init.c void sync_icache_dcache(pte_t pte) pte 79 arch/hexagon/mm/init.c page = pte_page(pte); pte 176 arch/ia64/include/asm/page.h typedef struct { unsigned long pte; } pte_t; pte 185 arch/ia64/include/asm/page.h # define pte_val(x) ((x).pte) pte 75 arch/ia64/include/asm/pgalloc.h pmd_populate(struct mm_struct *mm, pmd_t * pmd_entry, pgtable_t pte) pte 77 arch/ia64/include/asm/pgalloc.h pmd_val(*pmd_entry) = page_to_phys(pte); pte 82 arch/ia64/include/asm/pgalloc.h pmd_populate_kernel(struct mm_struct *mm, pmd_t * pmd_entry, pte_t * pte) pte 84 arch/ia64/include/asm/pgalloc.h pmd_val(*pmd_entry) = __pa(pte); pte 87 arch/ia64/include/asm/pgalloc.h #define __pte_free_tlb(tlb, pte, address) pte_free((tlb)->mm, pte) pte 265 arch/ia64/include/asm/pgtable.h #define pte_none(pte) (!pte_val(pte)) pte 266 arch/ia64/include/asm/pgtable.h #define pte_present(pte) (pte_val(pte) & (_PAGE_P | _PAGE_PROTNONE)) pte 267 arch/ia64/include/asm/pgtable.h #define pte_clear(mm,addr,pte) (pte_val(*(pte)) = 0UL) pte 269 arch/ia64/include/asm/pgtable.h #define pte_page(pte) virt_to_page(((pte_val(pte) & _PFN_MASK) + PAGE_OFFSET)) pte 297 arch/ia64/include/asm/pgtable.h #define pte_write(pte) ((unsigned) (((pte_val(pte) & _PAGE_AR_MASK) >> _PAGE_AR_SHIFT) - 2) <= 4) pte 298 arch/ia64/include/asm/pgtable.h #define pte_exec(pte) ((pte_val(pte) & _PAGE_AR_RX) != 0) pte 299 arch/ia64/include/asm/pgtable.h #define pte_dirty(pte) ((pte_val(pte) & _PAGE_D) != 0) pte 300 arch/ia64/include/asm/pgtable.h #define pte_young(pte) ((pte_val(pte) & _PAGE_A) != 0) pte 301 arch/ia64/include/asm/pgtable.h #define pte_special(pte) 0 pte 307 arch/ia64/include/asm/pgtable.h #define pte_wrprotect(pte) (__pte(pte_val(pte) & ~_PAGE_AR_RW)) pte 308 arch/ia64/include/asm/pgtable.h #define pte_mkwrite(pte) (__pte(pte_val(pte) | _PAGE_AR_RW)) pte 309 arch/ia64/include/asm/pgtable.h #define pte_mkold(pte) (__pte(pte_val(pte) & ~_PAGE_A)) pte 310 arch/ia64/include/asm/pgtable.h #define pte_mkyoung(pte) (__pte(pte_val(pte) | _PAGE_A)) pte 311 arch/ia64/include/asm/pgtable.h #define pte_mkclean(pte) (__pte(pte_val(pte) & ~_PAGE_D)) pte 312 arch/ia64/include/asm/pgtable.h #define pte_mkdirty(pte) (__pte(pte_val(pte) | _PAGE_D)) pte 313 arch/ia64/include/asm/pgtable.h #define pte_mkhuge(pte) (__pte(pte_val(pte))) pte 314 arch/ia64/include/asm/pgtable.h #define pte_mkspecial(pte) (pte) pte 325 arch/ia64/include/asm/pgtable.h #define pte_present_exec_user(pte)\ pte 326 arch/ia64/include/asm/pgtable.h ((pte_val(pte) & (_PAGE_P | _PAGE_PL_MASK | _PAGE_AR_RX)) == \ pte 405 arch/ia64/include/asm/pgtable.h #define pte_unmap(pte) do { } while (0) pte 417 arch/ia64/include/asm/pgtable.h pte_t pte = *ptep; pte 418 arch/ia64/include/asm/pgtable.h if (!pte_young(pte)) pte 420 arch/ia64/include/asm/pgtable.h set_pte_at(vma->vm_mm, addr, ptep, pte_mkold(pte)); pte 431 arch/ia64/include/asm/pgtable.h pte_t pte = *ptep; pte 433 arch/ia64/include/asm/pgtable.h return pte; pte 479 arch/ia64/include/asm/pgtable.h #define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val(pte) }) pte 481 arch/ia64/include/asm/processor.h __u64 vmaddr, __u64 pte, pte 488 arch/ia64/include/asm/processor.h ia64_itri(tr_num, pte); pte 490 arch/ia64/include/asm/processor.h ia64_itrd(tr_num, pte); pte 498 arch/ia64/include/asm/processor.h ia64_itc (__u64 target_mask, __u64 vmaddr, __u64 pte, pte 506 arch/ia64/include/asm/processor.h ia64_itci(pte); pte 508 arch/ia64/include/asm/processor.h ia64_itcd(pte); pte 20 arch/ia64/include/asm/tlbflush.h u64 pte; pte 24 arch/ia64/include/asm/tlbflush.h extern int ia64_itr_entry(u64 target_mask, u64 va, u64 pte, u64 log_size); pte 1236 arch/ia64/kernel/mca.c if (p->pte & 0x1) { pte 1245 arch/ia64/kernel/mca.c ia64_itr(0x1, i, p->ifa, p->pte, p->itir >> 2); pte 1249 arch/ia64/kernel/mca.c ia64_itr(0x2, i, p->ifa, p->pte, p->itir >> 2); pte 34 arch/ia64/mm/fault.c pte_t *ptep, pte; pte 52 arch/ia64/mm/fault.c pte = *ptep; pte 53 arch/ia64/mm/fault.c return pte_present(pte); pte 35 arch/ia64/mm/hugetlbpage.c pte_t *pte = NULL; pte 42 arch/ia64/mm/hugetlbpage.c pte = pte_alloc_map(mm, pmd, taddr); pte 44 arch/ia64/mm/hugetlbpage.c return pte; pte 54 arch/ia64/mm/hugetlbpage.c pte_t *pte = NULL; pte 62 arch/ia64/mm/hugetlbpage.c pte = pte_offset_map(pmd, taddr); pte 66 arch/ia64/mm/hugetlbpage.c return pte; pte 56 arch/ia64/mm/init.c __ia64_sync_icache_dcache (pte_t pte) pte 61 arch/ia64/mm/init.c page = pte_page(pte); pte 213 arch/ia64/mm/init.c pte_t *pte; pte 224 arch/ia64/mm/init.c pte = pte_alloc_kernel(pmd, address); pte 225 arch/ia64/mm/init.c if (!pte) pte 227 arch/ia64/mm/init.c if (!pte_none(*pte)) pte 229 arch/ia64/mm/init.c set_pte(pte, mk_pte(page, pgprot)); pte 387 arch/ia64/mm/init.c pte_t *pte; pte 407 arch/ia64/mm/init.c pte = pte_offset_kernel(pmd, end_address); pte 409 arch/ia64/mm/init.c if (pte_none(*pte)) { pte 411 arch/ia64/mm/init.c pte++; pte 435 arch/ia64/mm/init.c pte_t *pte; pte 463 arch/ia64/mm/init.c pte = memblock_alloc_node(PAGE_SIZE, PAGE_SIZE, node); pte 464 arch/ia64/mm/init.c if (!pte) pte 466 arch/ia64/mm/init.c pmd_populate_kernel(&init_mm, pmd, pte); pte 468 arch/ia64/mm/init.c pte = pte_offset_kernel(pmd, address); pte 470 arch/ia64/mm/init.c if (pte_none(*pte)) { pte 475 arch/ia64/mm/init.c set_pte(pte, pfn_pte(__pa(page) >> PAGE_SHIFT, pte 455 arch/ia64/mm/tlb.c int ia64_itr_entry(u64 target_mask, u64 va, u64 pte, u64 log_size) pte 475 arch/ia64/mm/tlb.c if (p->pte & 0x1) pte 487 arch/ia64/mm/tlb.c if (p->pte & 0x1) pte 499 arch/ia64/mm/tlb.c if (!((ia64_idtrs[cpu] + i)->pte & 0x1)) pte 503 arch/ia64/mm/tlb.c if (!((ia64_idtrs[cpu] + IA64_TR_ALLOC_MAX + i)->pte & 0x1)) pte 507 arch/ia64/mm/tlb.c if (!((ia64_idtrs[cpu] + i)->pte & 0x1) && pte 508 arch/ia64/mm/tlb.c !((ia64_idtrs[cpu] + IA64_TR_ALLOC_MAX + i)->pte & 0x1)) pte 526 arch/ia64/mm/tlb.c ia64_itr(0x1, i, va, pte, log_size); pte 530 arch/ia64/mm/tlb.c p->pte = pte; pte 535 arch/ia64/mm/tlb.c ia64_itr(0x2, i, va, pte, log_size); pte 539 arch/ia64/mm/tlb.c p->pte = pte; pte 569 arch/ia64/mm/tlb.c if ((p->pte&0x1) && is_tr_overlap(p, p->ifa, p->itir>>2)) { pte 570 arch/ia64/mm/tlb.c p->pte = 0; pte 578 arch/ia64/mm/tlb.c if ((p->pte & 0x1) && is_tr_overlap(p, p->ifa, p->itir>>2)) { pte 579 arch/ia64/mm/tlb.c p->pte = 0; pte 586 arch/ia64/mm/tlb.c if (((ia64_idtrs[cpu] + i)->pte & 0x1) || pte 587 arch/ia64/mm/tlb.c ((ia64_idtrs[cpu] + IA64_TR_ALLOC_MAX + i)->pte & 0x1)) pte 8 arch/m68k/include/asm/mcf_pgalloc.h extern inline void pte_free_kernel(struct mm_struct *mm, pte_t *pte) pte 10 arch/m68k/include/asm/mcf_pgalloc.h free_page((unsigned long) pte); pte 37 arch/m68k/include/asm/mcf_pgalloc.h #define pmd_populate_kernel(mm, pmd, pte) (pmd_val(*pmd) = (unsigned long)(pte)) pte 53 arch/m68k/include/asm/mcf_pgalloc.h pte_t *pte; pte 62 arch/m68k/include/asm/mcf_pgalloc.h pte = kmap(page); pte 63 arch/m68k/include/asm/mcf_pgalloc.h if (pte) { pte 64 arch/m68k/include/asm/mcf_pgalloc.h clear_page(pte); pte 65 arch/m68k/include/asm/mcf_pgalloc.h __flush_page_to_ram(pte); pte 66 arch/m68k/include/asm/mcf_pgalloc.h flush_tlb_kernel_page(pte); pte 67 arch/m68k/include/asm/mcf_pgalloc.h nocache_page(pte); pte 103 arch/m68k/include/asm/mcf_pgalloc.h #define pgd_populate(mm, pmd, pte) BUG() pte 159 arch/m68k/include/asm/mcf_pgtable.h static inline pte_t pte_modify(pte_t pte, pgprot_t newprot) pte 161 arch/m68k/include/asm/mcf_pgtable.h pte_val(pte) = (pte_val(pte) & CF_PAGE_CHG_MASK) | pgprot_val(newprot); pte 162 arch/m68k/include/asm/mcf_pgtable.h return pte; pte 172 arch/m68k/include/asm/mcf_pgtable.h #define __pte_page(pte) ((unsigned long) (pte_val(pte) & PAGE_MASK)) pte 175 arch/m68k/include/asm/mcf_pgtable.h static inline int pte_none(pte_t pte) pte 177 arch/m68k/include/asm/mcf_pgtable.h return !pte_val(pte); pte 180 arch/m68k/include/asm/mcf_pgtable.h static inline int pte_present(pte_t pte) pte 182 arch/m68k/include/asm/mcf_pgtable.h return pte_val(pte) & CF_PAGE_VALID; pte 191 arch/m68k/include/asm/mcf_pgtable.h #define pte_pagenr(pte) ((__pte_page(pte) - PAGE_OFFSET) >> PAGE_SHIFT) pte 192 arch/m68k/include/asm/mcf_pgtable.h #define pte_page(pte) virt_to_page(__pte_page(pte)) pte 221 arch/m68k/include/asm/mcf_pgtable.h static inline int pte_read(pte_t pte) pte 223 arch/m68k/include/asm/mcf_pgtable.h return pte_val(pte) & CF_PAGE_READABLE; pte 226 arch/m68k/include/asm/mcf_pgtable.h static inline int pte_write(pte_t pte) pte 228 arch/m68k/include/asm/mcf_pgtable.h return pte_val(pte) & CF_PAGE_WRITABLE; pte 231 arch/m68k/include/asm/mcf_pgtable.h static inline int pte_exec(pte_t pte) pte 233 arch/m68k/include/asm/mcf_pgtable.h return pte_val(pte) & CF_PAGE_EXEC; pte 236 arch/m68k/include/asm/mcf_pgtable.h static inline int pte_dirty(pte_t pte) pte 238 arch/m68k/include/asm/mcf_pgtable.h return pte_val(pte) & CF_PAGE_DIRTY; pte 241 arch/m68k/include/asm/mcf_pgtable.h static inline int pte_young(pte_t pte) pte 243 arch/m68k/include/asm/mcf_pgtable.h return pte_val(pte) & CF_PAGE_ACCESSED; pte 246 arch/m68k/include/asm/mcf_pgtable.h static inline int pte_special(pte_t pte) pte 251 arch/m68k/include/asm/mcf_pgtable.h static inline pte_t pte_wrprotect(pte_t pte) pte 253 arch/m68k/include/asm/mcf_pgtable.h pte_val(pte) &= ~CF_PAGE_WRITABLE; pte 254 arch/m68k/include/asm/mcf_pgtable.h return pte; pte 257 arch/m68k/include/asm/mcf_pgtable.h static inline pte_t pte_rdprotect(pte_t pte) pte 259 arch/m68k/include/asm/mcf_pgtable.h pte_val(pte) &= ~CF_PAGE_READABLE; pte 260 arch/m68k/include/asm/mcf_pgtable.h return pte; pte 263 arch/m68k/include/asm/mcf_pgtable.h static inline pte_t pte_exprotect(pte_t pte) pte 265 arch/m68k/include/asm/mcf_pgtable.h pte_val(pte) &= ~CF_PAGE_EXEC; pte 266 arch/m68k/include/asm/mcf_pgtable.h return pte; pte 269 arch/m68k/include/asm/mcf_pgtable.h static inline pte_t pte_mkclean(pte_t pte) pte 271 arch/m68k/include/asm/mcf_pgtable.h pte_val(pte) &= ~CF_PAGE_DIRTY; pte 272 arch/m68k/include/asm/mcf_pgtable.h return pte; pte 275 arch/m68k/include/asm/mcf_pgtable.h static inline pte_t pte_mkold(pte_t pte) pte 277 arch/m68k/include/asm/mcf_pgtable.h pte_val(pte) &= ~CF_PAGE_ACCESSED; pte 278 arch/m68k/include/asm/mcf_pgtable.h return pte; pte 281 arch/m68k/include/asm/mcf_pgtable.h static inline pte_t pte_mkwrite(pte_t pte) pte 283 arch/m68k/include/asm/mcf_pgtable.h pte_val(pte) |= CF_PAGE_WRITABLE; pte 284 arch/m68k/include/asm/mcf_pgtable.h return pte; pte 287 arch/m68k/include/asm/mcf_pgtable.h static inline pte_t pte_mkread(pte_t pte) pte 289 arch/m68k/include/asm/mcf_pgtable.h pte_val(pte) |= CF_PAGE_READABLE; pte 290 arch/m68k/include/asm/mcf_pgtable.h return pte; pte 293 arch/m68k/include/asm/mcf_pgtable.h static inline pte_t pte_mkexec(pte_t pte) pte 295 arch/m68k/include/asm/mcf_pgtable.h pte_val(pte) |= CF_PAGE_EXEC; pte 296 arch/m68k/include/asm/mcf_pgtable.h return pte; pte 299 arch/m68k/include/asm/mcf_pgtable.h static inline pte_t pte_mkdirty(pte_t pte) pte 301 arch/m68k/include/asm/mcf_pgtable.h pte_val(pte) |= CF_PAGE_DIRTY; pte 302 arch/m68k/include/asm/mcf_pgtable.h return pte; pte 305 arch/m68k/include/asm/mcf_pgtable.h static inline pte_t pte_mkyoung(pte_t pte) pte 307 arch/m68k/include/asm/mcf_pgtable.h pte_val(pte) |= CF_PAGE_ACCESSED; pte 308 arch/m68k/include/asm/mcf_pgtable.h return pte; pte 311 arch/m68k/include/asm/mcf_pgtable.h static inline pte_t pte_mknocache(pte_t pte) pte 313 arch/m68k/include/asm/mcf_pgtable.h pte_val(pte) |= 0x80 | (pte_val(pte) & ~0x40); pte 314 arch/m68k/include/asm/mcf_pgtable.h return pte; pte 317 arch/m68k/include/asm/mcf_pgtable.h static inline pte_t pte_mkcache(pte_t pte) pte 319 arch/m68k/include/asm/mcf_pgtable.h pte_val(pte) &= ~CF_PAGE_NOCACHE; pte 320 arch/m68k/include/asm/mcf_pgtable.h return pte; pte 323 arch/m68k/include/asm/mcf_pgtable.h static inline pte_t pte_mkspecial(pte_t pte) pte 325 arch/m68k/include/asm/mcf_pgtable.h return pte; pte 396 arch/m68k/include/asm/mcf_pgtable.h #define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val(pte) }) pte 403 arch/m68k/include/asm/mcf_pgtable.h #define pte_unmap(pte) ((void) 0) pte 405 arch/m68k/include/asm/mcf_pgtable.h #define pte_pfn(pte) (pte_val(pte) >> PAGE_SHIFT) pte 104 arch/m68k/include/asm/mmu_context.h pte_t *pte; pte 134 arch/m68k/include/asm/mmu_context.h pte = (mmuar >= PAGE_OFFSET) ? pte_offset_kernel(pmd, mmuar) pte 136 arch/m68k/include/asm/mmu_context.h if (pte_none(*pte) || !pte_present(*pte)) pte 139 arch/m68k/include/asm/mmu_context.h set_pte(pte, pte_mkyoung(*pte)); pte 141 arch/m68k/include/asm/mmu_context.h if (!pte_dirty(*pte) && mmuar <= PAGE_OFFSET) pte 142 arch/m68k/include/asm/mmu_context.h set_pte(pte, pte_wrprotect(*pte)); pte 145 arch/m68k/include/asm/mmu_context.h (((int)(pte->pte) & (int)CF_PAGE_MMUTR_MASK) pte 148 arch/m68k/include/asm/mmu_context.h mmu_write(MMUDR, (pte_val(*pte) & PAGE_MASK) | pte 149 arch/m68k/include/asm/mmu_context.h ((pte->pte) & CF_PAGE_MMUDR_MASK) | MMUDR_SZ_8KB | MMUDR_X); pte 13 arch/m68k/include/asm/motorola_pgalloc.h pte_t *pte; pte 15 arch/m68k/include/asm/motorola_pgalloc.h pte = (pte_t *)__get_free_page(GFP_KERNEL|__GFP_ZERO); pte 16 arch/m68k/include/asm/motorola_pgalloc.h if (pte) { pte 17 arch/m68k/include/asm/motorola_pgalloc.h __flush_page_to_ram(pte); pte 18 arch/m68k/include/asm/motorola_pgalloc.h flush_tlb_kernel_page(pte); pte 19 arch/m68k/include/asm/motorola_pgalloc.h nocache_page(pte); pte 22 arch/m68k/include/asm/motorola_pgalloc.h return pte; pte 25 arch/m68k/include/asm/motorola_pgalloc.h static inline void pte_free_kernel(struct mm_struct *mm, pte_t *pte) pte 27 arch/m68k/include/asm/motorola_pgalloc.h cache_page(pte); pte 28 arch/m68k/include/asm/motorola_pgalloc.h free_page((unsigned long) pte); pte 34 arch/m68k/include/asm/motorola_pgalloc.h pte_t *pte; pte 44 arch/m68k/include/asm/motorola_pgalloc.h pte = kmap(page); pte 45 arch/m68k/include/asm/motorola_pgalloc.h __flush_page_to_ram(pte); pte 46 arch/m68k/include/asm/motorola_pgalloc.h flush_tlb_kernel_page(pte); pte 47 arch/m68k/include/asm/motorola_pgalloc.h nocache_page(pte); pte 98 arch/m68k/include/asm/motorola_pgalloc.h static inline void pmd_populate_kernel(struct mm_struct *mm, pmd_t *pmd, pte_t *pte) pte 100 arch/m68k/include/asm/motorola_pgalloc.h pmd_set(pmd, pte); pte 103 arch/m68k/include/asm/motorola_pgtable.h static inline pte_t pte_modify(pte_t pte, pgprot_t newprot) pte 105 arch/m68k/include/asm/motorola_pgtable.h pte_val(pte) = (pte_val(pte) & _PAGE_CHG_MASK) | pgprot_val(newprot); pte 106 arch/m68k/include/asm/motorola_pgtable.h return pte; pte 125 arch/m68k/include/asm/motorola_pgtable.h #define __pte_page(pte) ((unsigned long)__va(pte_val(pte) & PAGE_MASK)) pte 130 arch/m68k/include/asm/motorola_pgtable.h #define pte_none(pte) (!pte_val(pte)) pte 131 arch/m68k/include/asm/motorola_pgtable.h #define pte_present(pte) (pte_val(pte) & (_PAGE_PRESENT | _PAGE_PROTNONE)) pte 134 arch/m68k/include/asm/motorola_pgtable.h #define pte_page(pte) virt_to_page(__va(pte_val(pte))) pte 135 arch/m68k/include/asm/motorola_pgtable.h #define pte_pfn(pte) (pte_val(pte) >> PAGE_SHIFT) pte 168 arch/m68k/include/asm/motorola_pgtable.h static inline int pte_write(pte_t pte) { return !(pte_val(pte) & _PAGE_RONLY); } pte 169 arch/m68k/include/asm/motorola_pgtable.h static inline int pte_dirty(pte_t pte) { return pte_val(pte) & _PAGE_DIRTY; } pte 170 arch/m68k/include/asm/motorola_pgtable.h static inline int pte_young(pte_t pte) { return pte_val(pte) & _PAGE_ACCESSED; } pte 171 arch/m68k/include/asm/motorola_pgtable.h static inline int pte_special(pte_t pte) { return 0; } pte 173 arch/m68k/include/asm/motorola_pgtable.h static inline pte_t pte_wrprotect(pte_t pte) { pte_val(pte) |= _PAGE_RONLY; return pte; } pte 174 arch/m68k/include/asm/motorola_pgtable.h static inline pte_t pte_mkclean(pte_t pte) { pte_val(pte) &= ~_PAGE_DIRTY; return pte; } pte 175 arch/m68k/include/asm/motorola_pgtable.h static inline pte_t pte_mkold(pte_t pte) { pte_val(pte) &= ~_PAGE_ACCESSED; return pte; } pte 176 arch/m68k/include/asm/motorola_pgtable.h static inline pte_t pte_mkwrite(pte_t pte) { pte_val(pte) &= ~_PAGE_RONLY; return pte; } pte 177 arch/m68k/include/asm/motorola_pgtable.h static inline pte_t pte_mkdirty(pte_t pte) { pte_val(pte) |= _PAGE_DIRTY; return pte; } pte 178 arch/m68k/include/asm/motorola_pgtable.h static inline pte_t pte_mkyoung(pte_t pte) { pte_val(pte) |= _PAGE_ACCESSED; return pte; } pte 179 arch/m68k/include/asm/motorola_pgtable.h static inline pte_t pte_mknocache(pte_t pte) pte 181 arch/m68k/include/asm/motorola_pgtable.h pte_val(pte) = (pte_val(pte) & _CACHEMASK040) | m68k_pgtable_cachemode; pte 182 arch/m68k/include/asm/motorola_pgtable.h return pte; pte 184 arch/m68k/include/asm/motorola_pgtable.h static inline pte_t pte_mkcache(pte_t pte) pte 186 arch/m68k/include/asm/motorola_pgtable.h pte_val(pte) = (pte_val(pte) & _CACHEMASK040) | m68k_supervisor_cachemode; pte 187 arch/m68k/include/asm/motorola_pgtable.h return pte; pte 189 arch/m68k/include/asm/motorola_pgtable.h static inline pte_t pte_mkspecial(pte_t pte) { return pte; } pte 224 arch/m68k/include/asm/motorola_pgtable.h #define pte_unmap(pte) ((void)0) pte 272 arch/m68k/include/asm/motorola_pgtable.h #define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val(pte) }) pte 24 arch/m68k/include/asm/page.h typedef struct { unsigned long pte; } pte_t; pte 30 arch/m68k/include/asm/page.h #define pte_val(x) ((x).pte) pte 38 arch/m68k/include/asm/pgtable_no.h #define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val(pte) }) pte 22 arch/m68k/include/asm/sun3_pgalloc.h #define __pte_free_tlb(tlb,pte,addr) \ pte 24 arch/m68k/include/asm/sun3_pgalloc.h pgtable_pte_page_dtor(pte); \ pte 25 arch/m68k/include/asm/sun3_pgalloc.h tlb_remove_page((tlb), pte); \ pte 28 arch/m68k/include/asm/sun3_pgalloc.h static inline void pmd_populate_kernel(struct mm_struct *mm, pmd_t *pmd, pte_t *pte) pte 30 arch/m68k/include/asm/sun3_pgalloc.h pmd_val(*pmd) = __pa((unsigned long)pte); pte 61 arch/m68k/include/asm/sun3_pgalloc.h #define pgd_populate(mm, pmd, pte) BUG() pte 105 arch/m68k/include/asm/sun3_pgtable.h static inline pte_t pte_modify(pte_t pte, pgprot_t newprot) pte 107 arch/m68k/include/asm/sun3_pgtable.h pte_val(pte) = (pte_val(pte) & SUN3_PAGE_CHG_MASK) | pgprot_val(newprot); pte 108 arch/m68k/include/asm/sun3_pgtable.h return pte; pte 118 arch/m68k/include/asm/sun3_pgtable.h #define __pte_page(pte) \ pte 119 arch/m68k/include/asm/sun3_pgtable.h ((unsigned long) __va ((pte_val (pte) & SUN3_PAGE_PGNUM_MASK) << PAGE_SHIFT)) pte 123 arch/m68k/include/asm/sun3_pgtable.h static inline int pte_none (pte_t pte) { return !pte_val (pte); } pte 124 arch/m68k/include/asm/sun3_pgtable.h static inline int pte_present (pte_t pte) { return pte_val (pte) & SUN3_PAGE_VALID; } pte 130 arch/m68k/include/asm/sun3_pgtable.h #define pte_pfn(pte) (pte_val(pte) & SUN3_PAGE_PGNUM_MASK) pte 134 arch/m68k/include/asm/sun3_pgtable.h #define pte_page(pte) virt_to_page(__pte_page(pte)) pte 167 arch/m68k/include/asm/sun3_pgtable.h static inline int pte_write(pte_t pte) { return pte_val(pte) & SUN3_PAGE_WRITEABLE; } pte 168 arch/m68k/include/asm/sun3_pgtable.h static inline int pte_dirty(pte_t pte) { return pte_val(pte) & SUN3_PAGE_MODIFIED; } pte 169 arch/m68k/include/asm/sun3_pgtable.h static inline int pte_young(pte_t pte) { return pte_val(pte) & SUN3_PAGE_ACCESSED; } pte 170 arch/m68k/include/asm/sun3_pgtable.h static inline int pte_special(pte_t pte) { return 0; } pte 172 arch/m68k/include/asm/sun3_pgtable.h static inline pte_t pte_wrprotect(pte_t pte) { pte_val(pte) &= ~SUN3_PAGE_WRITEABLE; return pte; } pte 173 arch/m68k/include/asm/sun3_pgtable.h static inline pte_t pte_mkclean(pte_t pte) { pte_val(pte) &= ~SUN3_PAGE_MODIFIED; return pte; } pte 174 arch/m68k/include/asm/sun3_pgtable.h static inline pte_t pte_mkold(pte_t pte) { pte_val(pte) &= ~SUN3_PAGE_ACCESSED; return pte; } pte 175 arch/m68k/include/asm/sun3_pgtable.h static inline pte_t pte_mkwrite(pte_t pte) { pte_val(pte) |= SUN3_PAGE_WRITEABLE; return pte; } pte 176 arch/m68k/include/asm/sun3_pgtable.h static inline pte_t pte_mkdirty(pte_t pte) { pte_val(pte) |= SUN3_PAGE_MODIFIED; return pte; } pte 177 arch/m68k/include/asm/sun3_pgtable.h static inline pte_t pte_mkyoung(pte_t pte) { pte_val(pte) |= SUN3_PAGE_ACCESSED; return pte; } pte 178 arch/m68k/include/asm/sun3_pgtable.h static inline pte_t pte_mknocache(pte_t pte) { pte_val(pte) |= SUN3_PAGE_NOCACHE; return pte; } pte 182 arch/m68k/include/asm/sun3_pgtable.h static inline pte_t pte_mkcache(pte_t pte) { return pte; } pte 183 arch/m68k/include/asm/sun3_pgtable.h static inline pte_t pte_mkspecial(pte_t pte) { return pte; } pte 207 arch/m68k/include/asm/sun3_pgtable.h #define pte_unmap(pte) do { } while (0) pte 213 arch/m68k/include/asm/sun3_pgtable.h #define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val(pte) }) pte 469 arch/m68k/kernel/sys_m68k.c pte_t *pte; pte 480 arch/m68k/kernel/sys_m68k.c pte = pte_offset_map_lock(mm, pmd, (unsigned long)mem, &ptl); pte 481 arch/m68k/kernel/sys_m68k.c if (!pte_present(*pte) || !pte_dirty(*pte) pte 482 arch/m68k/kernel/sys_m68k.c || !pte_write(*pte)) { pte 483 arch/m68k/kernel/sys_m68k.c pte_unmap_unlock(pte, ptl); pte 495 arch/m68k/kernel/sys_m68k.c pte_unmap_unlock(pte, ptl); pte 73 arch/m68k/mm/mcfmmu.c pte_t pte = pfn_pte(virt_to_pfn(address), PAGE_INIT); pte 75 arch/m68k/mm/mcfmmu.c pte_val(pte) = 0; pte 77 arch/m68k/mm/mcfmmu.c set_pte(pg_table, pte); pte 96 arch/m68k/mm/mcfmmu.c pte_t *pte; pte 122 arch/m68k/mm/mcfmmu.c pte = (KMAPAREA(mmuar)) ? pte_offset_kernel(pmd, mmuar) pte 124 arch/m68k/mm/mcfmmu.c if (pte_none(*pte) || !pte_present(*pte)) { pte 130 arch/m68k/mm/mcfmmu.c if (!pte_write(*pte)) { pte 134 arch/m68k/mm/mcfmmu.c set_pte(pte, pte_mkdirty(*pte)); pte 137 arch/m68k/mm/mcfmmu.c set_pte(pte, pte_mkyoung(*pte)); pte 139 arch/m68k/mm/mcfmmu.c if (!pte_dirty(*pte) && !KMAPAREA(mmuar)) pte 140 arch/m68k/mm/mcfmmu.c set_pte(pte, pte_wrprotect(*pte)); pte 144 arch/m68k/mm/mcfmmu.c mmutr |= (pte->pte & CF_PAGE_MMUTR_MASK) >> CF_PAGE_MMUTR_SHIFT; pte 147 arch/m68k/mm/mcfmmu.c mmu_write(MMUDR, (pte_val(*pte) & PAGE_MASK) | pte 148 arch/m68k/mm/mcfmmu.c ((pte->pte) & CF_PAGE_MMUDR_MASK) | MMUDR_SZ_8KB | MMUDR_X); pte 33 arch/m68k/mm/sun3kmap.c unsigned long pte; pte 37 arch/m68k/mm/sun3kmap.c pte = pte_val(ptep); pte 38 arch/m68k/mm/sun3kmap.c pte |= type; pte 40 arch/m68k/mm/sun3kmap.c sun3_put_pte(virt, pte); pte 79 arch/m68k/mm/sun3mmu.c pte_t pte = pfn_pte(virt_to_pfn(address), PAGE_INIT); pte 81 arch/m68k/mm/sun3mmu.c pte_val (pte) = 0; pte 82 arch/m68k/mm/sun3mmu.c set_pte (pg_table, pte); pte 26 arch/m68k/sun3/dvma.c unsigned long pte; pte 34 arch/m68k/sun3/dvma.c pte = pte_val(ptep); pte 36 arch/m68k/sun3/dvma.c if(ptelist[(vaddr & 0xff000) >> PAGE_SHIFT] != pte) { pte 37 arch/m68k/sun3/dvma.c sun3_put_pte(vaddr, pte); pte 38 arch/m68k/sun3/dvma.c ptelist[(vaddr & 0xff000) >> PAGE_SHIFT] = pte; pte 72 arch/m68k/sun3/mmu_emu.c void print_pte (pte_t pte) pte 76 arch/m68k/sun3/mmu_emu.c unsigned long val = pte_val (pte); pte 94 arch/m68k/sun3/mmu_emu.c unsigned long val = pte_val (pte); pte 360 arch/m68k/sun3/mmu_emu.c pte_t *pte; pte 387 arch/m68k/sun3/mmu_emu.c pte = (pte_t *) pgd_val (*(crp + segment)); pte 390 arch/m68k/sun3/mmu_emu.c if (!pte) { pte 395 arch/m68k/sun3/mmu_emu.c pte = (pte_t *) __va ((unsigned long)(pte + offset)); pte 398 arch/m68k/sun3/mmu_emu.c if (!(pte_val (*pte) & SUN3_PAGE_VALID)) pte 406 arch/m68k/sun3/mmu_emu.c sun3_put_pte (vaddr&PAGE_MASK, pte_val (*pte)); pte 414 arch/m68k/sun3/mmu_emu.c if (pte_val (*pte) & SUN3_PAGE_WRITEABLE) pte 415 arch/m68k/sun3/mmu_emu.c pte_val (*pte) |= (SUN3_PAGE_ACCESSED pte 420 arch/m68k/sun3/mmu_emu.c pte_val (*pte) |= SUN3_PAGE_ACCESSED; pte 109 arch/m68k/sun3x/dvma.c pte_t *pte; pte 112 arch/m68k/sun3x/dvma.c if((pte = pte_alloc_kernel(pmd, vaddr)) == NULL) { pte 125 arch/m68k/sun3x/dvma.c set_pte(pte, pfn_pte(virt_to_pfn(kaddr), pte 127 arch/m68k/sun3x/dvma.c pte++; pte 89 arch/microblaze/include/asm/page.h typedef struct { unsigned long pte; } pte_t; pte 102 arch/microblaze/include/asm/page.h # define pte_val(x) ((x).pte) pte 53 arch/microblaze/include/asm/pgalloc.h #define __pte_free_tlb(tlb, pte, addr) pte_free((tlb)->mm, (pte)) pte 55 arch/microblaze/include/asm/pgalloc.h #define pmd_populate(mm, pmd, pte) \ pte 56 arch/microblaze/include/asm/pgalloc.h (pmd_val(*(pmd)) = (unsigned long)page_address(pte)) pte 58 arch/microblaze/include/asm/pgalloc.h #define pmd_populate_kernel(mm, pmd, pte) \ pte 59 arch/microblaze/include/asm/pgalloc.h (pmd_val(*(pmd)) = (unsigned long) (pte)) pte 68 arch/microblaze/include/asm/pgalloc.h #define pgd_populate(mm, pmd, pte) BUG() pte 42 arch/microblaze/include/asm/pgtable.h #define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val(pte) }) pte 85 arch/microblaze/include/asm/pgtable.h static inline int pte_special(pte_t pte) { return 0; } pte 87 arch/microblaze/include/asm/pgtable.h static inline pte_t pte_mkspecial(pte_t pte) { return pte; } pte 296 arch/microblaze/include/asm/pgtable.h #define pte_none(pte) ((pte_val(pte) & ~_PTE_NONE_MASK) == 0) pte 297 arch/microblaze/include/asm/pgtable.h #define pte_present(pte) (pte_val(pte) & _PAGE_PRESENT) pte 332 arch/microblaze/include/asm/pgtable.h static inline int pte_read(pte_t pte) { return pte_val(pte) & _PAGE_USER; } pte 333 arch/microblaze/include/asm/pgtable.h static inline int pte_write(pte_t pte) { return pte_val(pte) & _PAGE_RW; } pte 334 arch/microblaze/include/asm/pgtable.h static inline int pte_exec(pte_t pte) { return pte_val(pte) & _PAGE_EXEC; } pte 335 arch/microblaze/include/asm/pgtable.h static inline int pte_dirty(pte_t pte) { return pte_val(pte) & _PAGE_DIRTY; } pte 336 arch/microblaze/include/asm/pgtable.h static inline int pte_young(pte_t pte) { return pte_val(pte) & _PAGE_ACCESSED; } pte 338 arch/microblaze/include/asm/pgtable.h static inline void pte_uncache(pte_t pte) { pte_val(pte) |= _PAGE_NO_CACHE; } pte 339 arch/microblaze/include/asm/pgtable.h static inline void pte_cache(pte_t pte) { pte_val(pte) &= ~_PAGE_NO_CACHE; } pte 341 arch/microblaze/include/asm/pgtable.h static inline pte_t pte_rdprotect(pte_t pte) \ pte 342 arch/microblaze/include/asm/pgtable.h { pte_val(pte) &= ~_PAGE_USER; return pte; } pte 343 arch/microblaze/include/asm/pgtable.h static inline pte_t pte_wrprotect(pte_t pte) \ pte 344 arch/microblaze/include/asm/pgtable.h { pte_val(pte) &= ~(_PAGE_RW | _PAGE_HWWRITE); return pte; } pte 345 arch/microblaze/include/asm/pgtable.h static inline pte_t pte_exprotect(pte_t pte) \ pte 346 arch/microblaze/include/asm/pgtable.h { pte_val(pte) &= ~_PAGE_EXEC; return pte; } pte 347 arch/microblaze/include/asm/pgtable.h static inline pte_t pte_mkclean(pte_t pte) \ pte 348 arch/microblaze/include/asm/pgtable.h { pte_val(pte) &= ~(_PAGE_DIRTY | _PAGE_HWWRITE); return pte; } pte 349 arch/microblaze/include/asm/pgtable.h static inline pte_t pte_mkold(pte_t pte) \ pte 350 arch/microblaze/include/asm/pgtable.h { pte_val(pte) &= ~_PAGE_ACCESSED; return pte; } pte 352 arch/microblaze/include/asm/pgtable.h static inline pte_t pte_mkread(pte_t pte) \ pte 353 arch/microblaze/include/asm/pgtable.h { pte_val(pte) |= _PAGE_USER; return pte; } pte 354 arch/microblaze/include/asm/pgtable.h static inline pte_t pte_mkexec(pte_t pte) \ pte 355 arch/microblaze/include/asm/pgtable.h { pte_val(pte) |= _PAGE_USER | _PAGE_EXEC; return pte; } pte 356 arch/microblaze/include/asm/pgtable.h static inline pte_t pte_mkwrite(pte_t pte) \ pte 357 arch/microblaze/include/asm/pgtable.h { pte_val(pte) |= _PAGE_RW; return pte; } pte 358 arch/microblaze/include/asm/pgtable.h static inline pte_t pte_mkdirty(pte_t pte) \ pte 359 arch/microblaze/include/asm/pgtable.h { pte_val(pte) |= _PAGE_DIRTY; return pte; } pte 360 arch/microblaze/include/asm/pgtable.h static inline pte_t pte_mkyoung(pte_t pte) \ pte 361 arch/microblaze/include/asm/pgtable.h { pte_val(pte) |= _PAGE_ACCESSED; return pte; } pte 370 arch/microblaze/include/asm/pgtable.h pte_t pte; pte 371 arch/microblaze/include/asm/pgtable.h pte_val(pte) = physpage | pgprot_val(pgprot); pte 372 arch/microblaze/include/asm/pgtable.h return pte; pte 377 arch/microblaze/include/asm/pgtable.h pte_t pte; \ pte 378 arch/microblaze/include/asm/pgtable.h pte_val(pte) = (((page - mem_map) << PAGE_SHIFT) + memory_start) | \ pte 380 arch/microblaze/include/asm/pgtable.h pte; \ pte 383 arch/microblaze/include/asm/pgtable.h static inline pte_t pte_modify(pte_t pte, pgprot_t newprot) pte 385 arch/microblaze/include/asm/pgtable.h pte_val(pte) = (pte_val(pte) & _PAGE_CHG_MASK) | pgprot_val(newprot); pte 386 arch/microblaze/include/asm/pgtable.h return pte; pte 421 arch/microblaze/include/asm/pgtable.h pte_t *ptep, pte_t pte) pte 423 arch/microblaze/include/asm/pgtable.h *ptep = pte; pte 427 arch/microblaze/include/asm/pgtable.h pte_t *ptep, pte_t pte) pte 429 arch/microblaze/include/asm/pgtable.h *ptep = pte; pte 496 arch/microblaze/include/asm/pgtable.h #define pte_unmap(pte) kunmap_atomic(pte) pte 510 arch/microblaze/include/asm/pgtable.h #define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val(pte) >> 2 }) pte 192 arch/microblaze/mm/pgtable.c pte_t *pte; pte 199 arch/microblaze/mm/pgtable.c pte = pte_offset_kernel(pmd, addr & PAGE_MASK); pte 200 arch/microblaze/mm/pgtable.c if (pte) { pte 202 arch/microblaze/mm/pgtable.c *ptep = pte; pte 216 arch/microblaze/mm/pgtable.c pte_t *pte; pte 228 arch/microblaze/mm/pgtable.c if (get_pteptr(mm, addr, &pte)) pte 229 arch/microblaze/mm/pgtable.c pa = (pte_val(*pte) & PAGE_MASK) | (addr & ~PAGE_MASK); pte 236 arch/microblaze/mm/pgtable.c pte_t *pte; pte 238 arch/microblaze/mm/pgtable.c pte = (pte_t *)__get_free_page(GFP_KERNEL | __GFP_ZERO); pte 240 arch/microblaze/mm/pgtable.c pte = (pte_t *)early_get_page(); pte 241 arch/microblaze/mm/pgtable.c if (pte) pte 242 arch/microblaze/mm/pgtable.c clear_page(pte); pte 244 arch/microblaze/mm/pgtable.c return pte; pte 45 arch/mips/include/asm/hugetlb.h pte_t pte = *ptep; pte 49 arch/mips/include/asm/hugetlb.h return pte; pte 60 arch/mips/include/asm/hugetlb.h static inline int huge_pte_none(pte_t pte) pte 62 arch/mips/include/asm/hugetlb.h unsigned long val = pte_val(pte) & ~_PAGE_GLOBAL; pte 69 arch/mips/include/asm/hugetlb.h pte_t *ptep, pte_t pte, pte 72 arch/mips/include/asm/hugetlb.h int changed = !pte_same(*ptep, pte); pte 75 arch/mips/include/asm/hugetlb.h set_pte_at(vma->vm_mm, addr, ptep, pte); pte 943 arch/mips/include/asm/kvm_host.h int kvm_set_spte_hva(struct kvm *kvm, unsigned long hva, pte_t pte); pte 128 arch/mips/include/asm/page.h typedef struct { unsigned long long pte; } pte_t; pte 129 arch/mips/include/asm/page.h #define pte_val(x) ((x).pte) pte 133 arch/mips/include/asm/page.h typedef struct { unsigned long pte; } pte_t; pte 134 arch/mips/include/asm/page.h #define pte_val(x) ((x).pte) pte 19 arch/mips/include/asm/pgalloc.h pte_t *pte) pte 21 arch/mips/include/asm/pgalloc.h set_pmd(pmd, __pmd((unsigned long)pte)); pte 25 arch/mips/include/asm/pgalloc.h pgtable_t pte) pte 27 arch/mips/include/asm/pgalloc.h set_pmd(pmd, __pmd((unsigned long)page_address(pte))); pte 55 arch/mips/include/asm/pgalloc.h #define __pte_free_tlb(tlb,pte,address) \ pte 57 arch/mips/include/asm/pgalloc.h pgtable_pte_page_dtor(pte); \ pte 58 arch/mips/include/asm/pgalloc.h tlb_remove_page((tlb), pte); \ pte 162 arch/mips/include/asm/pgtable-32.h pte_t pte; pte 164 arch/mips/include/asm/pgtable-32.h pte.pte_low = (pfn >> _PAGE_PRESENT_SHIFT) | pte 166 arch/mips/include/asm/pgtable-32.h pte.pte_high = (pfn << _PFN_SHIFT) | pte 168 arch/mips/include/asm/pgtable-32.h return pte; pte 177 arch/mips/include/asm/pgtable-32.h pte_t pte; pte 179 arch/mips/include/asm/pgtable-32.h pte.pte_high = (pfn << 6) | (pgprot_val(prot) & 0x3f); pte 180 arch/mips/include/asm/pgtable-32.h pte.pte_low = pgprot_val(prot); pte 182 arch/mips/include/asm/pgtable-32.h return pte; pte 188 arch/mips/include/asm/pgtable-32.h #define pte_pfn(x) ((unsigned long)((x).pte >> (PAGE_SHIFT + 2))) pte 191 arch/mips/include/asm/pgtable-32.h #define pte_pfn(x) ((unsigned long)((x).pte >> _PFN_SHIFT)) pte 222 arch/mips/include/asm/pgtable-32.h #define pte_unmap(pte) ((void)(pte)) pte 230 arch/mips/include/asm/pgtable-32.h #define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val(pte) }) pte 241 arch/mips/include/asm/pgtable-32.h #define __pte_to_swp_entry(pte) ((swp_entry_t) { (pte).pte_high }) pte 250 arch/mips/include/asm/pgtable-32.h #define __pte_to_swp_entry(pte) ((swp_entry_t) { (pte).pte_high }) pte 264 arch/mips/include/asm/pgtable-32.h #define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val(pte) }) pte 314 arch/mips/include/asm/pgtable-64.h #define pte_pfn(x) ((unsigned long)((x).pte >> (PAGE_SHIFT + 2))) pte 317 arch/mips/include/asm/pgtable-64.h #define pte_pfn(x) ((unsigned long)((x).pte >> _PFN_SHIFT)) pte 359 arch/mips/include/asm/pgtable-64.h #define pte_unmap(pte) ((void)(pte)) pte 373 arch/mips/include/asm/pgtable-64.h { pte_t pte; pte_val(pte) = (type << 16) | (offset << 24); return pte; } pte 378 arch/mips/include/asm/pgtable-64.h #define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val(pte) }) pte 139 arch/mips/include/asm/pgtable.h # define pte_none(pte) (!(((pte).pte_high) & ~_PAGE_GLOBAL)) pte 141 arch/mips/include/asm/pgtable.h # define pte_none(pte) (!(((pte).pte_low | (pte).pte_high) & ~_PAGE_GLOBAL)) pte 144 arch/mips/include/asm/pgtable.h #define pte_present(pte) ((pte).pte_low & _PAGE_PRESENT) pte 145 arch/mips/include/asm/pgtable.h #define pte_no_exec(pte) ((pte).pte_low & _PAGE_NO_EXEC) pte 147 arch/mips/include/asm/pgtable.h static inline void set_pte(pte_t *ptep, pte_t pte) pte 149 arch/mips/include/asm/pgtable.h ptep->pte_high = pte.pte_high; pte 151 arch/mips/include/asm/pgtable.h ptep->pte_low = pte.pte_low; pte 154 arch/mips/include/asm/pgtable.h if (pte.pte_high & _PAGE_GLOBAL) { pte 156 arch/mips/include/asm/pgtable.h if (pte.pte_low & _PAGE_GLOBAL) { pte 190 arch/mips/include/asm/pgtable.h #define pte_none(pte) (!(pte_val(pte) & ~_PAGE_GLOBAL)) pte 191 arch/mips/include/asm/pgtable.h #define pte_present(pte) (pte_val(pte) & _PAGE_PRESENT) pte 192 arch/mips/include/asm/pgtable.h #define pte_no_exec(pte) (pte_val(pte) & _PAGE_NO_EXEC) pte 210 arch/mips/include/asm/pgtable.h cmpxchg64(&buddy->pte, 0, _PAGE_GLOBAL); pte 212 arch/mips/include/asm/pgtable.h cmpxchg(&buddy->pte, 0, _PAGE_GLOBAL); pte 235 arch/mips/include/asm/pgtable.h extern void __update_cache(unsigned long address, pte_t pte); pte 277 arch/mips/include/asm/pgtable.h static inline int pte_write(pte_t pte) { return pte.pte_low & _PAGE_WRITE; } pte 278 arch/mips/include/asm/pgtable.h static inline int pte_dirty(pte_t pte) { return pte.pte_low & _PAGE_MODIFIED; } pte 279 arch/mips/include/asm/pgtable.h static inline int pte_young(pte_t pte) { return pte.pte_low & _PAGE_ACCESSED; } pte 280 arch/mips/include/asm/pgtable.h static inline int pte_special(pte_t pte) { return pte.pte_low & _PAGE_SPECIAL; } pte 282 arch/mips/include/asm/pgtable.h static inline pte_t pte_wrprotect(pte_t pte) pte 284 arch/mips/include/asm/pgtable.h pte.pte_low &= ~_PAGE_WRITE; pte 286 arch/mips/include/asm/pgtable.h pte.pte_low &= ~_PAGE_SILENT_WRITE; pte 287 arch/mips/include/asm/pgtable.h pte.pte_high &= ~_PAGE_SILENT_WRITE; pte 288 arch/mips/include/asm/pgtable.h return pte; pte 291 arch/mips/include/asm/pgtable.h static inline pte_t pte_mkclean(pte_t pte) pte 293 arch/mips/include/asm/pgtable.h pte.pte_low &= ~_PAGE_MODIFIED; pte 295 arch/mips/include/asm/pgtable.h pte.pte_low &= ~_PAGE_SILENT_WRITE; pte 296 arch/mips/include/asm/pgtable.h pte.pte_high &= ~_PAGE_SILENT_WRITE; pte 297 arch/mips/include/asm/pgtable.h return pte; pte 300 arch/mips/include/asm/pgtable.h static inline pte_t pte_mkold(pte_t pte) pte 302 arch/mips/include/asm/pgtable.h pte.pte_low &= ~_PAGE_ACCESSED; pte 304 arch/mips/include/asm/pgtable.h pte.pte_low &= ~_PAGE_SILENT_READ; pte 305 arch/mips/include/asm/pgtable.h pte.pte_high &= ~_PAGE_SILENT_READ; pte 306 arch/mips/include/asm/pgtable.h return pte; pte 309 arch/mips/include/asm/pgtable.h static inline pte_t pte_mkwrite(pte_t pte) pte 311 arch/mips/include/asm/pgtable.h pte.pte_low |= _PAGE_WRITE; pte 312 arch/mips/include/asm/pgtable.h if (pte.pte_low & _PAGE_MODIFIED) { pte 314 arch/mips/include/asm/pgtable.h pte.pte_low |= _PAGE_SILENT_WRITE; pte 315 arch/mips/include/asm/pgtable.h pte.pte_high |= _PAGE_SILENT_WRITE; pte 317 arch/mips/include/asm/pgtable.h return pte; pte 320 arch/mips/include/asm/pgtable.h static inline pte_t pte_mkdirty(pte_t pte) pte 322 arch/mips/include/asm/pgtable.h pte.pte_low |= _PAGE_MODIFIED; pte 323 arch/mips/include/asm/pgtable.h if (pte.pte_low & _PAGE_WRITE) { pte 325 arch/mips/include/asm/pgtable.h pte.pte_low |= _PAGE_SILENT_WRITE; pte 326 arch/mips/include/asm/pgtable.h pte.pte_high |= _PAGE_SILENT_WRITE; pte 328 arch/mips/include/asm/pgtable.h return pte; pte 331 arch/mips/include/asm/pgtable.h static inline pte_t pte_mkyoung(pte_t pte) pte 333 arch/mips/include/asm/pgtable.h pte.pte_low |= _PAGE_ACCESSED; pte 334 arch/mips/include/asm/pgtable.h if (!(pte.pte_low & _PAGE_NO_READ)) { pte 336 arch/mips/include/asm/pgtable.h pte.pte_low |= _PAGE_SILENT_READ; pte 337 arch/mips/include/asm/pgtable.h pte.pte_high |= _PAGE_SILENT_READ; pte 339 arch/mips/include/asm/pgtable.h return pte; pte 342 arch/mips/include/asm/pgtable.h static inline pte_t pte_mkspecial(pte_t pte) pte 344 arch/mips/include/asm/pgtable.h pte.pte_low |= _PAGE_SPECIAL; pte 345 arch/mips/include/asm/pgtable.h return pte; pte 348 arch/mips/include/asm/pgtable.h static inline int pte_write(pte_t pte) { return pte_val(pte) & _PAGE_WRITE; } pte 349 arch/mips/include/asm/pgtable.h static inline int pte_dirty(pte_t pte) { return pte_val(pte) & _PAGE_MODIFIED; } pte 350 arch/mips/include/asm/pgtable.h static inline int pte_young(pte_t pte) { return pte_val(pte) & _PAGE_ACCESSED; } pte 351 arch/mips/include/asm/pgtable.h static inline int pte_special(pte_t pte) { return pte_val(pte) & _PAGE_SPECIAL; } pte 353 arch/mips/include/asm/pgtable.h static inline pte_t pte_wrprotect(pte_t pte) pte 355 arch/mips/include/asm/pgtable.h pte_val(pte) &= ~(_PAGE_WRITE | _PAGE_SILENT_WRITE); pte 356 arch/mips/include/asm/pgtable.h return pte; pte 359 arch/mips/include/asm/pgtable.h static inline pte_t pte_mkclean(pte_t pte) pte 361 arch/mips/include/asm/pgtable.h pte_val(pte) &= ~(_PAGE_MODIFIED | _PAGE_SILENT_WRITE); pte 362 arch/mips/include/asm/pgtable.h return pte; pte 365 arch/mips/include/asm/pgtable.h static inline pte_t pte_mkold(pte_t pte) pte 367 arch/mips/include/asm/pgtable.h pte_val(pte) &= ~(_PAGE_ACCESSED | _PAGE_SILENT_READ); pte 368 arch/mips/include/asm/pgtable.h return pte; pte 371 arch/mips/include/asm/pgtable.h static inline pte_t pte_mkwrite(pte_t pte) pte 373 arch/mips/include/asm/pgtable.h pte_val(pte) |= _PAGE_WRITE; pte 374 arch/mips/include/asm/pgtable.h if (pte_val(pte) & _PAGE_MODIFIED) pte 375 arch/mips/include/asm/pgtable.h pte_val(pte) |= _PAGE_SILENT_WRITE; pte 376 arch/mips/include/asm/pgtable.h return pte; pte 379 arch/mips/include/asm/pgtable.h static inline pte_t pte_mkdirty(pte_t pte) pte 381 arch/mips/include/asm/pgtable.h pte_val(pte) |= _PAGE_MODIFIED; pte 382 arch/mips/include/asm/pgtable.h if (pte_val(pte) & _PAGE_WRITE) pte 383 arch/mips/include/asm/pgtable.h pte_val(pte) |= _PAGE_SILENT_WRITE; pte 384 arch/mips/include/asm/pgtable.h return pte; pte 387 arch/mips/include/asm/pgtable.h static inline pte_t pte_mkyoung(pte_t pte) pte 389 arch/mips/include/asm/pgtable.h pte_val(pte) |= _PAGE_ACCESSED; pte 390 arch/mips/include/asm/pgtable.h if (!(pte_val(pte) & _PAGE_NO_READ)) pte 391 arch/mips/include/asm/pgtable.h pte_val(pte) |= _PAGE_SILENT_READ; pte 392 arch/mips/include/asm/pgtable.h return pte; pte 395 arch/mips/include/asm/pgtable.h static inline pte_t pte_mkspecial(pte_t pte) pte 397 arch/mips/include/asm/pgtable.h pte_val(pte) |= _PAGE_SPECIAL; pte 398 arch/mips/include/asm/pgtable.h return pte; pte 402 arch/mips/include/asm/pgtable.h static inline int pte_huge(pte_t pte) { return pte_val(pte) & _PAGE_HUGE; } pte 404 arch/mips/include/asm/pgtable.h static inline pte_t pte_mkhuge(pte_t pte) pte 406 arch/mips/include/asm/pgtable.h pte_val(pte) |= _PAGE_HUGE; pte 407 arch/mips/include/asm/pgtable.h return pte; pte 448 arch/mips/include/asm/pgtable.h static inline pte_t pte_modify(pte_t pte, pgprot_t newprot) pte 450 arch/mips/include/asm/pgtable.h pte.pte_low &= (_PAGE_MODIFIED | _PAGE_ACCESSED | _PFNX_MASK); pte 451 arch/mips/include/asm/pgtable.h pte.pte_high &= (_PFN_MASK | _CACHE_MASK); pte 452 arch/mips/include/asm/pgtable.h pte.pte_low |= pgprot_val(newprot) & ~_PFNX_MASK; pte 453 arch/mips/include/asm/pgtable.h pte.pte_high |= pgprot_val(newprot) & ~(_PFN_MASK | _CACHE_MASK); pte 454 arch/mips/include/asm/pgtable.h return pte; pte 457 arch/mips/include/asm/pgtable.h static inline pte_t pte_modify(pte_t pte, pgprot_t newprot) pte 459 arch/mips/include/asm/pgtable.h pte.pte_low &= _PAGE_CHG_MASK; pte 460 arch/mips/include/asm/pgtable.h pte.pte_high &= (_PFN_MASK | _CACHE_MASK); pte 461 arch/mips/include/asm/pgtable.h pte.pte_low |= pgprot_val(newprot); pte 462 arch/mips/include/asm/pgtable.h pte.pte_high |= pgprot_val(newprot) & ~(_PFN_MASK | _CACHE_MASK); pte 463 arch/mips/include/asm/pgtable.h return pte; pte 466 arch/mips/include/asm/pgtable.h static inline pte_t pte_modify(pte_t pte, pgprot_t newprot) pte 468 arch/mips/include/asm/pgtable.h return __pte((pte_val(pte) & _PAGE_CHG_MASK) | pte 475 arch/mips/include/asm/pgtable.h pte_t pte); pte 480 arch/mips/include/asm/pgtable.h pte_t pte = *ptep; pte 481 arch/mips/include/asm/pgtable.h __update_tlb(vma, address, pte); pte 487 arch/mips/include/asm/pgtable.h pte_t pte = *(pte_t *)pmdp; pte 489 arch/mips/include/asm/pgtable.h __update_tlb(vma, address, pte); pte 185 arch/mips/kvm/mmu.c static bool kvm_mips_flush_gpa_pte(pte_t *pte, unsigned long start_gpa, pte 194 arch/mips/kvm/mmu.c if (!pte_present(pte[i])) pte 197 arch/mips/kvm/mmu.c set_pte(pte + i, __pte(0)); pte 205 arch/mips/kvm/mmu.c pte_t *pte; pte 216 arch/mips/kvm/mmu.c pte = pte_offset(pmd + i, 0); pte 220 arch/mips/kvm/mmu.c if (kvm_mips_flush_gpa_pte(pte, start_gpa, end)) { pte 222 arch/mips/kvm/mmu.c pte_free_kernel(NULL, pte); pte 307 arch/mips/kvm/mmu.c static int kvm_mips_##name##_pte(pte_t *pte, unsigned long start, \ pte 317 arch/mips/kvm/mmu.c if (!pte_present(pte[i])) \ pte 320 arch/mips/kvm/mmu.c old = pte[i]; \ pte 324 arch/mips/kvm/mmu.c set_pte(pte + i, new); \ pte 335 arch/mips/kvm/mmu.c pte_t *pte; \ pte 345 arch/mips/kvm/mmu.c pte = pte_offset(pmd + i, 0); \ pte 349 arch/mips/kvm/mmu.c ret |= kvm_mips_##name##_pte(pte, start, cur_end); \ pte 554 arch/mips/kvm/mmu.c int kvm_set_spte_hva(struct kvm *kvm, unsigned long hva, pte_t pte) pte 559 arch/mips/kvm/mmu.c ret = handle_hva_to_gpa(kvm, hva, end, &kvm_set_spte_handler, &pte); pte 836 arch/mips/kvm/mmu.c static bool kvm_mips_flush_gva_pte(pte_t *pte, unsigned long start_gva, pte 852 arch/mips/kvm/mmu.c if (!pte_present(pte[i])) pte 855 arch/mips/kvm/mmu.c set_pte(pte + i, __pte(0)); pte 863 arch/mips/kvm/mmu.c pte_t *pte; pte 874 arch/mips/kvm/mmu.c pte = pte_offset(pmd + i, 0); pte 878 arch/mips/kvm/mmu.c if (kvm_mips_flush_gva_pte(pte, start_gva, end)) { pte 880 arch/mips/kvm/mmu.c pte_free_kernel(NULL, pte); pte 964 arch/mips/kvm/mmu.c static pte_t kvm_mips_gpa_pte_to_gva_unmapped(pte_t pte) pte 971 arch/mips/kvm/mmu.c if (!pte_dirty(pte)) pte 972 arch/mips/kvm/mmu.c pte = pte_wrprotect(pte); pte 974 arch/mips/kvm/mmu.c return pte; pte 977 arch/mips/kvm/mmu.c static pte_t kvm_mips_gpa_pte_to_gva_mapped(pte_t pte, long entrylo) pte 981 arch/mips/kvm/mmu.c pte = pte_mkclean(pte); pte 983 arch/mips/kvm/mmu.c return kvm_mips_gpa_pte_to_gva_unmapped(pte); pte 569 arch/mips/kvm/trap_emul.c pte_t *pte; pte 595 arch/mips/kvm/trap_emul.c pte = pte_offset(pmd + k, 0); pte 596 arch/mips/kvm/trap_emul.c pte_free_kernel(NULL, pte); pte 130 arch/mips/mm/cache.c void __update_cache(unsigned long address, pte_t pte) pte 134 arch/mips/mm/cache.c int exec = !pte_no_exec(pte) && !cpu_has_ic_fills_f_dc; pte 136 arch/mips/mm/cache.c pfn = pte_pfn(pte); pte 29 arch/mips/mm/hugetlbpage.c pte_t *pte = NULL; pte 34 arch/mips/mm/hugetlbpage.c pte = (pte_t *)pmd_alloc(mm, pud, addr); pte 36 arch/mips/mm/hugetlbpage.c return pte; pte 90 arch/mips/mm/init.c pte_t pte; pte 100 arch/mips/mm/init.c pte = mk_pte(page, prot); pte 102 arch/mips/mm/init.c entrylo = pte_to_entrylo(pte.pte_high); pte 104 arch/mips/mm/init.c entrylo = pte.pte_high; pte 106 arch/mips/mm/init.c entrylo = pte_to_entrylo(pte_val(pte)); pte 120 arch/mips/mm/init.c entrylo = (pte.pte_low & _PFNX_MASK); pte 237 arch/mips/mm/init.c pte_t *pte; pte 253 arch/mips/mm/init.c pte = (pte_t *) memblock_alloc_low(PAGE_SIZE, pte 255 arch/mips/mm/init.c if (!pte) pte 260 arch/mips/mm/init.c set_pmd(pmd, __pmd((unsigned long)pte)); pte 261 arch/mips/mm/init.c BUG_ON(pte != pte_offset_kernel(pmd, 0)); pte 21 arch/mips/mm/ioremap.c static inline void remap_area_pte(pte_t * pte, unsigned long address, pte 36 arch/mips/mm/ioremap.c if (!pte_none(*pte)) { pte 40 arch/mips/mm/ioremap.c set_pte(pte, pfn_pte(pfn, pgprot)); pte 43 arch/mips/mm/ioremap.c pte++; pte 59 arch/mips/mm/ioremap.c pte_t * pte = pte_alloc_kernel(pmd, address); pte 60 arch/mips/mm/ioremap.c if (!pte) pte 62 arch/mips/mm/ioremap.c remap_area_pte(pte, address, end - address, address + phys_addr, flags); pte 61 arch/mips/mm/pgtable-32.c pte_t *pte; pte 87 arch/mips/mm/pgtable-32.c pte = pte_offset_kernel(pmd, vaddr); pte 88 arch/mips/mm/pgtable-32.c pkmap_page_table = pte; pte 182 arch/mips/mm/tlb-r3k.c void __update_tlb(struct vm_area_struct *vma, unsigned long address, pte_t pte) pte 209 arch/mips/mm/tlb-r3k.c write_c0_entrylo0(pte_val(pte)); pte 294 arch/mips/mm/tlb-r4k.c void __update_tlb(struct vm_area_struct * vma, unsigned long address, pte_t pte) pte 731 arch/mips/mm/tlbex.c static void build_huge_update_entries(u32 **p, unsigned int pte, pte 751 arch/mips/mm/tlbex.c build_convert_pte_to_entrylo(p, pte); pte 752 arch/mips/mm/tlbex.c UASM_i_MTC0(p, pte, C0_ENTRYLO0); /* load it */ pte 755 arch/mips/mm/tlbex.c UASM_i_ADDIU(p, pte, pte, HPAGE_SIZE >> 7); pte 757 arch/mips/mm/tlbex.c UASM_i_ADDU(p, pte, pte, tmp); pte 759 arch/mips/mm/tlbex.c UASM_i_MTC0(p, pte, C0_ENTRYLO1); /* load it */ pte 764 arch/mips/mm/tlbex.c unsigned int pte, pte 769 arch/mips/mm/tlbex.c UASM_i_SC(p, pte, 0, ptr); pte 770 arch/mips/mm/tlbex.c uasm_il_beqz(p, r, pte, label_tlb_huge_update); pte 771 arch/mips/mm/tlbex.c UASM_i_LW(p, pte, 0, ptr); /* Needed because SC killed our PTE */ pte 773 arch/mips/mm/tlbex.c UASM_i_SW(p, pte, 0, ptr); pte 785 arch/mips/mm/tlbex.c build_huge_update_entries(p, pte, ptr); pte 786 arch/mips/mm/tlbex.c build_huge_tlb_write_entry(p, l, r, pte, tlb_random, 0); pte 791 arch/mips/mm/tlbex.c build_huge_update_entries(p, pte, ptr); pte 792 arch/mips/mm/tlbex.c build_huge_tlb_write_entry(p, l, r, pte, tlb_indexed, 0); pte 1667 arch/mips/mm/tlbex.c iPTE_LW(u32 **p, unsigned int pte, unsigned int ptr) pte 1674 arch/mips/mm/tlbex.c uasm_i_lld(p, pte, 0, ptr); pte 1677 arch/mips/mm/tlbex.c UASM_i_LL(p, pte, 0, ptr); pte 1681 arch/mips/mm/tlbex.c uasm_i_ld(p, pte, 0, ptr); pte 1684 arch/mips/mm/tlbex.c UASM_i_LW(p, pte, 0, ptr); pte 1689 arch/mips/mm/tlbex.c iPTE_SW(u32 **p, struct uasm_reloc **r, unsigned int pte, unsigned int ptr, pte 1697 arch/mips/mm/tlbex.c uasm_i_or(p, pte, pte, scratch); pte 1700 arch/mips/mm/tlbex.c uasm_i_ori(p, pte, pte, mode); pte 1706 arch/mips/mm/tlbex.c uasm_i_scd(p, pte, 0, ptr); pte 1709 arch/mips/mm/tlbex.c UASM_i_SC(p, pte, 0, ptr); pte 1712 arch/mips/mm/tlbex.c uasm_il_beqzl(p, r, pte, label_smp_pgtable_change); pte 1714 arch/mips/mm/tlbex.c uasm_il_beqz(p, r, pte, label_smp_pgtable_change); pte 1719 arch/mips/mm/tlbex.c uasm_i_ll(p, pte, sizeof(pte_t) / 2, ptr); pte 1720 arch/mips/mm/tlbex.c uasm_i_ori(p, pte, pte, hwmode); pte 1722 arch/mips/mm/tlbex.c uasm_i_sc(p, pte, sizeof(pte_t) / 2, ptr); pte 1723 arch/mips/mm/tlbex.c uasm_il_beqz(p, r, pte, label_smp_pgtable_change); pte 1725 arch/mips/mm/tlbex.c uasm_i_lw(p, pte, 0, ptr); pte 1734 arch/mips/mm/tlbex.c uasm_i_sd(p, pte, 0, ptr); pte 1737 arch/mips/mm/tlbex.c UASM_i_SW(p, pte, 0, ptr); pte 1741 arch/mips/mm/tlbex.c uasm_i_lw(p, pte, sizeof(pte_t) / 2, ptr); pte 1742 arch/mips/mm/tlbex.c uasm_i_ori(p, pte, pte, hwmode); pte 1744 arch/mips/mm/tlbex.c uasm_i_sw(p, pte, sizeof(pte_t) / 2, ptr); pte 1745 arch/mips/mm/tlbex.c uasm_i_lw(p, pte, 0, ptr); pte 1758 arch/mips/mm/tlbex.c int pte, int ptr, int scratch, enum label_id lid) pte 1760 arch/mips/mm/tlbex.c int t = scratch >= 0 ? scratch : pte; pte 1761 arch/mips/mm/tlbex.c int cur = pte; pte 1765 arch/mips/mm/tlbex.c uasm_il_bbit0(p, r, pte, ilog2(_PAGE_PRESENT), lid); pte 1774 arch/mips/mm/tlbex.c if (pte == t) pte 1776 arch/mips/mm/tlbex.c iPTE_LW(p, pte, ptr); pte 1787 arch/mips/mm/tlbex.c if (pte == t) pte 1789 arch/mips/mm/tlbex.c iPTE_LW(p, pte, ptr); pte 1795 arch/mips/mm/tlbex.c build_make_valid(u32 **p, struct uasm_reloc **r, unsigned int pte, pte 1800 arch/mips/mm/tlbex.c iPTE_SW(p, r, pte, ptr, mode, scratch); pte 1809 arch/mips/mm/tlbex.c unsigned int pte, unsigned int ptr, int scratch, pte 1812 arch/mips/mm/tlbex.c int t = scratch >= 0 ? scratch : pte; pte 1813 arch/mips/mm/tlbex.c int cur = pte; pte 1824 arch/mips/mm/tlbex.c if (pte == t) pte 1826 arch/mips/mm/tlbex.c iPTE_LW(p, pte, ptr); pte 1835 arch/mips/mm/tlbex.c build_make_write(u32 **p, struct uasm_reloc **r, unsigned int pte, pte 1841 arch/mips/mm/tlbex.c iPTE_SW(p, r, pte, ptr, mode, scratch); pte 1850 arch/mips/mm/tlbex.c unsigned int pte, unsigned int ptr, int scratch, pte 1854 arch/mips/mm/tlbex.c uasm_il_bbit0(p, r, pte, ilog2(_PAGE_WRITE), lid); pte 1857 arch/mips/mm/tlbex.c int t = scratch >= 0 ? scratch : pte; pte 1858 arch/mips/mm/tlbex.c uasm_i_srl(p, t, pte, _PAGE_WRITE_SHIFT); pte 1861 arch/mips/mm/tlbex.c if (pte == t) pte 1863 arch/mips/mm/tlbex.c iPTE_LW(p, pte, ptr); pte 1879 arch/mips/mm/tlbex.c build_r3000_pte_reload_tlbwi(u32 **p, unsigned int pte, unsigned int tmp) pte 1881 arch/mips/mm/tlbex.c uasm_i_mtc0(p, pte, C0_ENTRYLO0); /* cp0 delay */ pte 1896 arch/mips/mm/tlbex.c struct uasm_reloc **r, unsigned int pte, pte 1900 arch/mips/mm/tlbex.c uasm_i_mtc0(p, pte, C0_ENTRYLO0); /* cp0 delay */ pte 1913 arch/mips/mm/tlbex.c build_r3000_tlbchange_handler_head(u32 **p, unsigned int pte, pte 1918 arch/mips/mm/tlbex.c uasm_i_mfc0(p, pte, C0_BADVADDR); pte 1921 arch/mips/mm/tlbex.c uasm_i_srl(p, pte, pte, 22); /* load delay */ pte 1922 arch/mips/mm/tlbex.c uasm_i_sll(p, pte, pte, 2); pte 1923 arch/mips/mm/tlbex.c uasm_i_addu(p, ptr, ptr, pte); pte 1924 arch/mips/mm/tlbex.c uasm_i_mfc0(p, pte, C0_CONTEXT); pte 1926 arch/mips/mm/tlbex.c uasm_i_andi(p, pte, pte, 0xffc); /* load delay */ pte 1927 arch/mips/mm/tlbex.c uasm_i_addu(p, ptr, ptr, pte); pte 1928 arch/mips/mm/tlbex.c uasm_i_lw(p, pte, 0, ptr); pte 316 arch/mips/sgi-ip22/ip28-berr.c unsigned long pte = (lo >> 6) << 12; /* PTEBase */ pte 317 arch/mips/sgi-ip22/ip28-berr.c pte += 8*((vaddr >> pgsz) & 0x1ff); pte 318 arch/mips/sgi-ip22/ip28-berr.c if (page_is_ram(PFN_DOWN(pte))) { pte 325 arch/mips/sgi-ip22/ip28-berr.c PHYS_TO_XKSEG_UNCACHED(pte); pte 20 arch/nds32/include/asm/pgalloc.h #define pgd_populate(mm, pmd, pte) BUG() pte 28 arch/nds32/include/asm/pgalloc.h pgtable_t pte; pte 30 arch/nds32/include/asm/pgalloc.h pte = __pte_alloc_one(mm, GFP_PGTABLE_USER); pte 31 arch/nds32/include/asm/pgalloc.h if (pte) pte 32 arch/nds32/include/asm/pgalloc.h cpu_dcache_wb_page((unsigned long)page_address(pte)); pte 34 arch/nds32/include/asm/pgalloc.h return pte; pte 40 arch/nds32/include/asm/pgtable.h #define pte_ERROR(pte) __pte_error(__FILE__, __LINE__, pte_val(pte)) pte 187 arch/nds32/include/asm/pgtable.h #define pte_pfn(pte) (pte_val(pte) >> PAGE_SHIFT) pte 190 arch/nds32/include/asm/pgtable.h #define pte_none(pte) !(pte_val(pte)) pte 192 arch/nds32/include/asm/pgtable.h #define pte_page(pte) (pfn_to_page(pte_pfn(pte))) pte 200 arch/nds32/include/asm/pgtable.h #define pte_unmap(pte) do { } while (0) pte 201 arch/nds32/include/asm/pgtable.h #define pte_unmap_nested(pte) do { } while (0) pte 224 arch/nds32/include/asm/pgtable.h static inline void set_pte(pte_t * ptep, pte_t pte) pte 227 arch/nds32/include/asm/pgtable.h *ptep = pte; pte 249 arch/nds32/include/asm/pgtable.h #define pte_present(pte) (pte_val(pte) & _PAGE_V) pte 250 arch/nds32/include/asm/pgtable.h #define pte_write(pte) ((pte_val(pte) & _PAGE_M_MASK) == _PAGE_M_URW_KRW) pte 251 arch/nds32/include/asm/pgtable.h #define pte_read(pte) (((pte_val(pte) & _PAGE_M_MASK) == _PAGE_M_UR_KR) || \ pte 252 arch/nds32/include/asm/pgtable.h ((pte_val(pte) & _PAGE_M_MASK) == _PAGE_M_UR_KRW) || \ pte 253 arch/nds32/include/asm/pgtable.h ((pte_val(pte) & _PAGE_M_MASK) == _PAGE_M_URW_KRW)) pte 254 arch/nds32/include/asm/pgtable.h #define pte_kernel_write(pte) (((pte_val(pte) & _PAGE_M_MASK) == _PAGE_M_URW_KRW) || \ pte 255 arch/nds32/include/asm/pgtable.h ((pte_val(pte) & _PAGE_M_MASK) == _PAGE_M_UR_KRW) || \ pte 256 arch/nds32/include/asm/pgtable.h ((pte_val(pte) & _PAGE_M_MASK) == _PAGE_M_KRW) || \ pte 257 arch/nds32/include/asm/pgtable.h (((pte_val(pte) & _PAGE_M_MASK) == _PAGE_M_XKRW) && pte_exec(pte))) pte 258 arch/nds32/include/asm/pgtable.h #define pte_exec(pte) (pte_val(pte) & _PAGE_E) pte 259 arch/nds32/include/asm/pgtable.h #define pte_dirty(pte) (pte_val(pte) & _PAGE_D) pte 260 arch/nds32/include/asm/pgtable.h #define pte_young(pte) (pte_val(pte) & _PAGE_YOUNG) pte 265 arch/nds32/include/asm/pgtable.h #define pte_file(pte) (pte_val(pte) & _PAGE_FILE) pte 272 arch/nds32/include/asm/pgtable.h static inline pte_t pte_##fn(pte_t pte) { pte_val(pte) op; return pte; } pte 274 arch/nds32/include/asm/pgtable.h static inline pte_t pte_wrprotect(pte_t pte) pte 276 arch/nds32/include/asm/pgtable.h pte_val(pte) = pte_val(pte) & ~_PAGE_M_MASK; pte 277 arch/nds32/include/asm/pgtable.h pte_val(pte) = pte_val(pte) | _PAGE_M_UR_KR; pte 278 arch/nds32/include/asm/pgtable.h return pte; pte 281 arch/nds32/include/asm/pgtable.h static inline pte_t pte_mkwrite(pte_t pte) pte 283 arch/nds32/include/asm/pgtable.h pte_val(pte) = pte_val(pte) & ~_PAGE_M_MASK; pte 284 arch/nds32/include/asm/pgtable.h pte_val(pte) = pte_val(pte) | _PAGE_M_URW_KRW; pte 285 arch/nds32/include/asm/pgtable.h return pte; pte 294 arch/nds32/include/asm/pgtable.h static inline int pte_special(pte_t pte) pte 299 arch/nds32/include/asm/pgtable.h static inline pte_t pte_mkspecial(pte_t pte) pte 301 arch/nds32/include/asm/pgtable.h return pte; pte 372 arch/nds32/include/asm/pgtable.h static inline pte_t pte_modify(pte_t pte, pgprot_t newprot) pte 375 arch/nds32/include/asm/pgtable.h pte_val(pte) = (pte_val(pte) & ~mask) | (pgprot_val(newprot) & mask); pte 376 arch/nds32/include/asm/pgtable.h return pte; pte 388 arch/nds32/include/asm/pgtable.h #define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val(pte) }) pte 9 arch/nds32/include/asm/tlb.h #define __pte_free_tlb(tlb, pte, addr) pte_free((tlb)->mm, pte) pte 44 arch/nds32/include/asm/tlbflush.h unsigned long address, pte_t * pte); pte 50 arch/nds32/mm/cacheflush.c pte_t * pte) pte 53 arch/nds32/mm/cacheflush.c unsigned long pfn = pte_pfn(*pte); pte 62 arch/nds32/mm/cacheflush.c __nds32__tlbop_rwr(*pte); pte 88 arch/nds32/mm/cacheflush.c unsigned long kaddr, pte; pte 92 arch/nds32/mm/cacheflush.c pte = (pa | PAGE_KERNEL); pte 94 arch/nds32/mm/cacheflush.c __nds32__tlbop_rwlk(pte); pte 108 arch/nds32/mm/cacheflush.c unsigned long kaddr, pte; pte 112 arch/nds32/mm/cacheflush.c pte = (pa | PAGE_KERNEL); pte 114 arch/nds32/mm/cacheflush.c __nds32__tlbop_rwlk(pte); pte 59 arch/nds32/mm/fault.c pte_t *pte; pte 61 arch/nds32/mm/fault.c pte = pte_offset_map(pmd, addr); pte 62 arch/nds32/mm/fault.c pr_alert(", *pte=%08lx", pte_val(*pte)); pte 63 arch/nds32/mm/fault.c pte_unmap(pte); pte 38 arch/nds32/mm/highmem.c unsigned long vaddr, pte; pte 51 arch/nds32/mm/highmem.c pte = (page_to_pfn(page) << PAGE_SHIFT) | (PAGE_KERNEL); pte 53 arch/nds32/mm/highmem.c set_pte(ptep, pte); pte 57 arch/nds32/mm/highmem.c __nds32__tlbop_rwr(pte); pte 59 arch/nds32/mm/init.c pte_t *pte; pte 81 arch/nds32/mm/init.c pte = memblock_alloc(PAGE_SIZE, PAGE_SIZE); pte 82 arch/nds32/mm/init.c if (!pte) pte 85 arch/nds32/mm/init.c set_pmd(pme, __pmd(__pa(pte) + _PAGE_KERNEL_TABLE)); pte 89 arch/nds32/mm/init.c v += PAGE_SIZE, p += PAGE_SIZE, j++, pte++) { pte 92 arch/nds32/mm/init.c set_pte(pte, __pte(p + pgprot_val(PAGE_KERNEL))); pte 106 arch/nds32/mm/init.c pte_t *pte; pte 131 arch/nds32/mm/init.c pte = memblock_alloc(PAGE_SIZE, PAGE_SIZE); pte 132 arch/nds32/mm/init.c if (!pte) pte 135 arch/nds32/mm/init.c set_pmd(pmd, __pmd(__pa(pte) + _PAGE_KERNEL_TABLE)); pte 136 arch/nds32/mm/init.c pkmap_page_table = pte; pte 259 arch/nds32/mm/init.c pte_t *pte; pte 263 arch/nds32/mm/init.c pte = (pte_t *)&fixmap_pmd_p[pte_index(addr)]; pte 266 arch/nds32/mm/init.c set_pte(pte, pfn_pte(phys >> PAGE_SHIFT, flags)); pte 268 arch/nds32/mm/init.c pte_clear(&init_mm, addr, pte); pte 44 arch/nds32/mm/mm-nds32.c struct page *pte; pte 58 arch/nds32/mm/mm-nds32.c pte = pmd_page(*pmd); pte 61 arch/nds32/mm/mm-nds32.c pte_free(mm, pte); pte 20 arch/nds32/mm/proc.c pte_t *ptep, pte; pte 25 arch/nds32/mm/proc.c pte = *ptep; pte 26 arch/nds32/mm/proc.c if (pte_present(pte)) pte 27 arch/nds32/mm/proc.c return pte; pte 37 arch/nds32/mm/proc.c pte_t *ptep, pte; pte 46 arch/nds32/mm/proc.c pte = *ptep; pte 47 arch/nds32/mm/proc.c if (pte_present(pte)) pte 48 arch/nds32/mm/proc.c return pte; pte 59 arch/nds32/mm/proc.c pte_t pte; pte 64 arch/nds32/mm/proc.c pte = va_present(mm, addr); pte 65 arch/nds32/mm/proc.c if (!pte && pte_read(pte)) pte 78 arch/nds32/mm/proc.c pte_t pte; pte 83 arch/nds32/mm/proc.c pte = va_present(mm, addr); pte 84 arch/nds32/mm/proc.c if (!pte && pte_write(pte)) pte 88 arch/nds32/mm/proc.c pte = va_kernel_present(addr); pte 89 arch/nds32/mm/proc.c if (!pte && pte_kernel_write(pte)) pte 61 arch/nios2/include/asm/page.h typedef struct { unsigned long pte; } pte_t; pte 65 arch/nios2/include/asm/page.h #define pte_val(x) ((x).pte) pte 18 arch/nios2/include/asm/pgalloc.h pte_t *pte) pte 20 arch/nios2/include/asm/pgalloc.h set_pmd(pmd, __pmd((unsigned long)pte)); pte 24 arch/nios2/include/asm/pgalloc.h pgtable_t pte) pte 26 arch/nios2/include/asm/pgalloc.h set_pmd(pmd, __pmd((unsigned long)page_address(pte))); pte 42 arch/nios2/include/asm/pgalloc.h #define __pte_free_tlb(tlb, pte, addr) \ pte 44 arch/nios2/include/asm/pgalloc.h pgtable_pte_page_dtor(pte); \ pte 45 arch/nios2/include/asm/pgalloc.h tlb_remove_page((tlb), (pte)); \ pte 110 arch/nios2/include/asm/pgtable.h static inline int pte_write(pte_t pte) \ pte 111 arch/nios2/include/asm/pgtable.h { return pte_val(pte) & _PAGE_WRITE; } pte 112 arch/nios2/include/asm/pgtable.h static inline int pte_dirty(pte_t pte) \ pte 113 arch/nios2/include/asm/pgtable.h { return pte_val(pte) & _PAGE_DIRTY; } pte 114 arch/nios2/include/asm/pgtable.h static inline int pte_young(pte_t pte) \ pte 115 arch/nios2/include/asm/pgtable.h { return pte_val(pte) & _PAGE_ACCESSED; } pte 116 arch/nios2/include/asm/pgtable.h static inline int pte_special(pte_t pte) { return 0; } pte 129 arch/nios2/include/asm/pgtable.h static inline int pte_none(pte_t pte) pte 131 arch/nios2/include/asm/pgtable.h return !(pte_val(pte) & ~(_PAGE_GLOBAL|0xf)); pte 134 arch/nios2/include/asm/pgtable.h static inline int pte_present(pte_t pte) \ pte 135 arch/nios2/include/asm/pgtable.h { return pte_val(pte) & _PAGE_PRESENT; } pte 141 arch/nios2/include/asm/pgtable.h static inline pte_t pte_wrprotect(pte_t pte) pte 143 arch/nios2/include/asm/pgtable.h pte_val(pte) &= ~_PAGE_WRITE; pte 144 arch/nios2/include/asm/pgtable.h return pte; pte 147 arch/nios2/include/asm/pgtable.h static inline pte_t pte_mkclean(pte_t pte) pte 149 arch/nios2/include/asm/pgtable.h pte_val(pte) &= ~_PAGE_DIRTY; pte 150 arch/nios2/include/asm/pgtable.h return pte; pte 153 arch/nios2/include/asm/pgtable.h static inline pte_t pte_mkold(pte_t pte) pte 155 arch/nios2/include/asm/pgtable.h pte_val(pte) &= ~_PAGE_ACCESSED; pte 156 arch/nios2/include/asm/pgtable.h return pte; pte 159 arch/nios2/include/asm/pgtable.h static inline pte_t pte_mkwrite(pte_t pte) pte 161 arch/nios2/include/asm/pgtable.h pte_val(pte) |= _PAGE_WRITE; pte 162 arch/nios2/include/asm/pgtable.h return pte; pte 165 arch/nios2/include/asm/pgtable.h static inline pte_t pte_mkdirty(pte_t pte) pte 167 arch/nios2/include/asm/pgtable.h pte_val(pte) |= _PAGE_DIRTY; pte 168 arch/nios2/include/asm/pgtable.h return pte; pte 171 arch/nios2/include/asm/pgtable.h static inline pte_t pte_mkspecial(pte_t pte) { return pte; } pte 173 arch/nios2/include/asm/pgtable.h static inline pte_t pte_mkyoung(pte_t pte) pte 175 arch/nios2/include/asm/pgtable.h pte_val(pte) |= _PAGE_ACCESSED; pte 176 arch/nios2/include/asm/pgtable.h return pte; pte 179 arch/nios2/include/asm/pgtable.h static inline pte_t pte_modify(pte_t pte, pgprot_t newprot) pte 183 arch/nios2/include/asm/pgtable.h pte_val(pte) = (pte_val(pte) & ~mask) | (pgprot_val(newprot) & mask); pte 184 arch/nios2/include/asm/pgtable.h return pte; pte 198 arch/nios2/include/asm/pgtable.h #define pte_pfn(pte) (pte_val(pte) & 0xfffff) pte 200 arch/nios2/include/asm/pgtable.h #define pte_page(pte) (pfn_to_page(pte_pfn(pte))) pte 243 arch/nios2/include/asm/pgtable.h #define pte_unmap(pte) do { } while (0) pte 288 arch/nios2/include/asm/pgtable.h #define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val(pte) }) pte 298 arch/nios2/include/asm/pgtable.h unsigned long address, pte_t *pte); pte 42 arch/nios2/include/asm/tlbflush.h pte_t pte); pte 203 arch/nios2/mm/cacheflush.c pte_t pte = *ptep; pte 204 arch/nios2/mm/cacheflush.c unsigned long pfn = pte_pfn(pte); pte 208 arch/nios2/mm/cacheflush.c reload_tlb_page(vma, address, pte); pte 22 arch/nios2/mm/ioremap.c static inline void remap_area_pte(pte_t *pte, unsigned long address, pte 39 arch/nios2/mm/ioremap.c if (!pte_none(*pte)) { pte 43 arch/nios2/mm/ioremap.c set_pte(pte, pfn_pte(pfn, pgprot)); pte 46 arch/nios2/mm/ioremap.c pte++; pte 64 arch/nios2/mm/ioremap.c pte_t *pte = pte_alloc_kernel(pmd, address); pte 66 arch/nios2/mm/ioremap.c if (!pte) pte 68 arch/nios2/mm/ioremap.c remap_area_pte(pte, address, end - address, address + phys_addr, pte 96 arch/nios2/mm/tlb.c static void reload_tlb_one_pid(unsigned long addr, unsigned long mmu_pid, pte_t pte) pte 100 arch/nios2/mm/tlb.c replace_tlb_one_pid(addr, mmu_pid, pte_val(pte)); pte 114 arch/nios2/mm/tlb.c void reload_tlb_page(struct vm_area_struct *vma, unsigned long addr, pte_t pte) pte 118 arch/nios2/mm/tlb.c reload_tlb_one_pid(addr, mmu_pid, pte); pte 49 arch/openrisc/include/asm/page.h unsigned long pte; pte 59 arch/openrisc/include/asm/page.h #define pte_val(x) ((x).pte) pte 25 arch/openrisc/include/asm/pgalloc.h #define pmd_populate_kernel(mm, pmd, pte) \ pte 26 arch/openrisc/include/asm/pgalloc.h set_pmd(pmd, __pmd(_KERNPG_TABLE + __pa(pte))) pte 29 arch/openrisc/include/asm/pgalloc.h struct page *pte) pte 32 arch/openrisc/include/asm/pgalloc.h ((unsigned long)page_to_pfn(pte) << pte 73 arch/openrisc/include/asm/pgalloc.h struct page *pte; pte 74 arch/openrisc/include/asm/pgalloc.h pte = alloc_pages(GFP_KERNEL, 0); pte 75 arch/openrisc/include/asm/pgalloc.h if (!pte) pte 77 arch/openrisc/include/asm/pgalloc.h clear_page(page_address(pte)); pte 78 arch/openrisc/include/asm/pgalloc.h if (!pgtable_pte_page_ctor(pte)) { pte 79 arch/openrisc/include/asm/pgalloc.h __free_page(pte); pte 82 arch/openrisc/include/asm/pgalloc.h return pte; pte 85 arch/openrisc/include/asm/pgalloc.h static inline void pte_free_kernel(struct mm_struct *mm, pte_t *pte) pte 87 arch/openrisc/include/asm/pgalloc.h free_page((unsigned long)pte); pte 90 arch/openrisc/include/asm/pgalloc.h static inline void pte_free(struct mm_struct *mm, struct page *pte) pte 92 arch/openrisc/include/asm/pgalloc.h pgtable_pte_page_dtor(pte); pte 93 arch/openrisc/include/asm/pgalloc.h __free_page(pte); pte 96 arch/openrisc/include/asm/pgalloc.h #define __pte_free_tlb(tlb, pte, addr) \ pte 98 arch/openrisc/include/asm/pgalloc.h pgtable_pte_page_dtor(pte); \ pte 99 arch/openrisc/include/asm/pgalloc.h tlb_remove_page((tlb), (pte)); \ pte 234 arch/openrisc/include/asm/pgtable.h static inline int pte_read(pte_t pte) { return pte_val(pte) & _PAGE_READ; } pte 235 arch/openrisc/include/asm/pgtable.h static inline int pte_write(pte_t pte) { return pte_val(pte) & _PAGE_WRITE; } pte 236 arch/openrisc/include/asm/pgtable.h static inline int pte_exec(pte_t pte) { return pte_val(pte) & _PAGE_EXEC; } pte 237 arch/openrisc/include/asm/pgtable.h static inline int pte_dirty(pte_t pte) { return pte_val(pte) & _PAGE_DIRTY; } pte 238 arch/openrisc/include/asm/pgtable.h static inline int pte_young(pte_t pte) { return pte_val(pte) & _PAGE_ACCESSED; } pte 239 arch/openrisc/include/asm/pgtable.h static inline int pte_special(pte_t pte) { return 0; } pte 240 arch/openrisc/include/asm/pgtable.h static inline pte_t pte_mkspecial(pte_t pte) { return pte; } pte 242 arch/openrisc/include/asm/pgtable.h static inline pte_t pte_wrprotect(pte_t pte) pte 244 arch/openrisc/include/asm/pgtable.h pte_val(pte) &= ~(_PAGE_WRITE); pte 245 arch/openrisc/include/asm/pgtable.h return pte; pte 248 arch/openrisc/include/asm/pgtable.h static inline pte_t pte_rdprotect(pte_t pte) pte 250 arch/openrisc/include/asm/pgtable.h pte_val(pte) &= ~(_PAGE_READ); pte 251 arch/openrisc/include/asm/pgtable.h return pte; pte 254 arch/openrisc/include/asm/pgtable.h static inline pte_t pte_exprotect(pte_t pte) pte 256 arch/openrisc/include/asm/pgtable.h pte_val(pte) &= ~(_PAGE_EXEC); pte 257 arch/openrisc/include/asm/pgtable.h return pte; pte 260 arch/openrisc/include/asm/pgtable.h static inline pte_t pte_mkclean(pte_t pte) pte 262 arch/openrisc/include/asm/pgtable.h pte_val(pte) &= ~(_PAGE_DIRTY); pte 263 arch/openrisc/include/asm/pgtable.h return pte; pte 266 arch/openrisc/include/asm/pgtable.h static inline pte_t pte_mkold(pte_t pte) pte 268 arch/openrisc/include/asm/pgtable.h pte_val(pte) &= ~(_PAGE_ACCESSED); pte 269 arch/openrisc/include/asm/pgtable.h return pte; pte 272 arch/openrisc/include/asm/pgtable.h static inline pte_t pte_mkwrite(pte_t pte) pte 274 arch/openrisc/include/asm/pgtable.h pte_val(pte) |= _PAGE_WRITE; pte 275 arch/openrisc/include/asm/pgtable.h return pte; pte 278 arch/openrisc/include/asm/pgtable.h static inline pte_t pte_mkread(pte_t pte) pte 280 arch/openrisc/include/asm/pgtable.h pte_val(pte) |= _PAGE_READ; pte 281 arch/openrisc/include/asm/pgtable.h return pte; pte 284 arch/openrisc/include/asm/pgtable.h static inline pte_t pte_mkexec(pte_t pte) pte 286 arch/openrisc/include/asm/pgtable.h pte_val(pte) |= _PAGE_EXEC; pte 287 arch/openrisc/include/asm/pgtable.h return pte; pte 290 arch/openrisc/include/asm/pgtable.h static inline pte_t pte_mkdirty(pte_t pte) pte 292 arch/openrisc/include/asm/pgtable.h pte_val(pte) |= _PAGE_DIRTY; pte 293 arch/openrisc/include/asm/pgtable.h return pte; pte 296 arch/openrisc/include/asm/pgtable.h static inline pte_t pte_mkyoung(pte_t pte) pte 298 arch/openrisc/include/asm/pgtable.h pte_val(pte) |= _PAGE_ACCESSED; pte 299 arch/openrisc/include/asm/pgtable.h return pte; pte 315 arch/openrisc/include/asm/pgtable.h pte_t pte; pte 317 arch/openrisc/include/asm/pgtable.h pte_val(pte) = __pa(page) | pgprot_val(pgprot); pte 318 arch/openrisc/include/asm/pgtable.h return pte; pte 331 arch/openrisc/include/asm/pgtable.h static inline pte_t pte_modify(pte_t pte, pgprot_t newprot) pte 333 arch/openrisc/include/asm/pgtable.h pte_val(pte) = (pte_val(pte) & _PAGE_CHG_MASK) | pgprot_val(newprot); pte 334 arch/openrisc/include/asm/pgtable.h return pte; pte 345 arch/openrisc/include/asm/pgtable.h static inline unsigned long __pte_page(pte_t pte) pte 348 arch/openrisc/include/asm/pgtable.h return (unsigned long)__va(pte_val(pte) & PAGE_MASK); pte 351 arch/openrisc/include/asm/pgtable.h #define pte_pagenr(pte) ((__pte_page(pte) - PAGE_OFFSET) >> PAGE_SHIFT) pte 356 arch/openrisc/include/asm/pgtable.h #define pte_page(pte) (mem_map+pte_pagenr(pte)) pte 399 arch/openrisc/include/asm/pgtable.h #define pte_unmap(pte) do { } while (0) pte 400 arch/openrisc/include/asm/pgtable.h #define pte_unmap_nested(pte) do { } while (0) pte 401 arch/openrisc/include/asm/pgtable.h #define pte_pfn(x) ((unsigned long)(((x).pte)) >> PAGE_SHIFT) pte 416 arch/openrisc/include/asm/pgtable.h unsigned long address, pte_t *pte) pte 421 arch/openrisc/include/asm/pgtable.h unsigned long address, pte_t *pte); pte 424 arch/openrisc/include/asm/pgtable.h unsigned long address, pte_t *pte) pte 426 arch/openrisc/include/asm/pgtable.h update_tlb(vma, address, pte); pte 427 arch/openrisc/include/asm/pgtable.h update_cache(vma, address, pte); pte 439 arch/openrisc/include/asm/pgtable.h #define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val(pte) }) pte 26 arch/openrisc/kernel/dma.c page_set_nocache(pte_t *pte, unsigned long addr, pte 32 arch/openrisc/kernel/dma.c pte_val(*pte) |= _PAGE_CI; pte 52 arch/openrisc/kernel/dma.c page_clear_nocache(pte_t *pte, unsigned long addr, pte 55 arch/openrisc/kernel/dma.c pte_val(*pte) &= ~_PAGE_CI; pte 43 arch/openrisc/mm/cache.c pte_t *pte) pte 45 arch/openrisc/mm/cache.c unsigned long pfn = pte_val(*pte) >> PAGE_SHIFT; pte 76 arch/openrisc/mm/init.c pte_t *pte; pte 103 arch/openrisc/mm/init.c pte = memblock_alloc_raw(PAGE_SIZE, PAGE_SIZE); pte 104 arch/openrisc/mm/init.c if (!pte) pte 107 arch/openrisc/mm/init.c set_pmd(pme, __pmd(_KERNPG_TABLE + __pa(pte))); pte 111 arch/openrisc/mm/init.c v += PAGE_SIZE, p += PAGE_SIZE, j++, pte++) { pte 118 arch/openrisc/mm/init.c set_pte(pte, mk_pte_phys(p, prot)); pte 119 arch/openrisc/mm/ioremap.c pte_t *pte; pte 122 arch/openrisc/mm/ioremap.c pte = (pte_t *)get_zeroed_page(GFP_KERNEL); pte 124 arch/openrisc/mm/ioremap.c pte = memblock_alloc(PAGE_SIZE, PAGE_SIZE); pte 125 arch/openrisc/mm/ioremap.c if (!pte) pte 130 arch/openrisc/mm/ioremap.c return pte; pte 9 arch/parisc/include/asm/hugetlb.h pte_t *ptep, pte_t pte); pte 49 arch/parisc/include/asm/hugetlb.h pte_t pte, int dirty); pte 41 arch/parisc/include/asm/page.h typedef struct { unsigned long pte; } pte_t; /* either 32 or 64bit */ pte 49 arch/parisc/include/asm/page.h #define pte_val(x) ((x).pte) pte 102 arch/parisc/include/asm/pgalloc.h #define pgd_populate(mm, pmd, pte) BUG() pte 107 arch/parisc/include/asm/pgalloc.h pmd_populate_kernel(struct mm_struct *mm, pmd_t *pmd, pte_t *pte) pte 116 arch/parisc/include/asm/pgalloc.h + (__u32)(__pa((unsigned long)pte) >> PxD_VALUE_SHIFT)); pte 120 arch/parisc/include/asm/pgalloc.h + (__u32)(__pa((unsigned long)pte) >> PxD_VALUE_SHIFT)); pte 379 arch/parisc/include/asm/pgtable.h static inline int pte_dirty(pte_t pte) { return pte_val(pte) & _PAGE_DIRTY; } pte 380 arch/parisc/include/asm/pgtable.h static inline int pte_young(pte_t pte) { return pte_val(pte) & _PAGE_ACCESSED; } pte 381 arch/parisc/include/asm/pgtable.h static inline int pte_write(pte_t pte) { return pte_val(pte) & _PAGE_WRITE; } pte 382 arch/parisc/include/asm/pgtable.h static inline int pte_special(pte_t pte) { return 0; } pte 384 arch/parisc/include/asm/pgtable.h static inline pte_t pte_mkclean(pte_t pte) { pte_val(pte) &= ~_PAGE_DIRTY; return pte; } pte 385 arch/parisc/include/asm/pgtable.h static inline pte_t pte_mkold(pte_t pte) { pte_val(pte) &= ~_PAGE_ACCESSED; return pte; } pte 386 arch/parisc/include/asm/pgtable.h static inline pte_t pte_wrprotect(pte_t pte) { pte_val(pte) &= ~_PAGE_WRITE; return pte; } pte 387 arch/parisc/include/asm/pgtable.h static inline pte_t pte_mkdirty(pte_t pte) { pte_val(pte) |= _PAGE_DIRTY; return pte; } pte 388 arch/parisc/include/asm/pgtable.h static inline pte_t pte_mkyoung(pte_t pte) { pte_val(pte) |= _PAGE_ACCESSED; return pte; } pte 389 arch/parisc/include/asm/pgtable.h static inline pte_t pte_mkwrite(pte_t pte) { pte_val(pte) |= _PAGE_WRITE; return pte; } pte 390 arch/parisc/include/asm/pgtable.h static inline pte_t pte_mkspecial(pte_t pte) { return pte; } pte 396 arch/parisc/include/asm/pgtable.h #define pte_huge(pte) (pte_val(pte) & _PAGE_HUGE) pte 397 arch/parisc/include/asm/pgtable.h #define pte_mkhuge(pte) (__pte(pte_val(pte) | \ pte 400 arch/parisc/include/asm/pgtable.h #define pte_huge(pte) (0) pte 401 arch/parisc/include/asm/pgtable.h #define pte_mkhuge(pte) (pte) pte 422 arch/parisc/include/asm/pgtable.h pte_t pte; pte 423 arch/parisc/include/asm/pgtable.h pte_val(pte) = (pfn << PFN_PTE_SHIFT) | pgprot_val(pgprot); pte 424 arch/parisc/include/asm/pgtable.h return pte; pte 427 arch/parisc/include/asm/pgtable.h static inline pte_t pte_modify(pte_t pte, pgprot_t newprot) pte 428 arch/parisc/include/asm/pgtable.h { pte_val(pte) = (pte_val(pte) & _PAGE_CHG_MASK) | pgprot_val(newprot); return pte; } pte 434 arch/parisc/include/asm/pgtable.h #define pte_page(pte) (pfn_to_page(pte_pfn(pte))) pte 465 arch/parisc/include/asm/pgtable.h #define pte_unmap(pte) do { } while (0) pte 467 arch/parisc/include/asm/pgtable.h #define pte_unmap(pte) do { } while (0) pte 468 arch/parisc/include/asm/pgtable.h #define pte_unmap_nested(pte) do { } while (0) pte 486 arch/parisc/include/asm/pgtable.h #define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val(pte) }) pte 500 arch/parisc/include/asm/pgtable.h pte_t pte; pte 507 arch/parisc/include/asm/pgtable.h pte = *ptep; pte 508 arch/parisc/include/asm/pgtable.h if (!pte_young(pte)) { pte 512 arch/parisc/include/asm/pgtable.h set_pte(ptep, pte_mkold(pte)); pte 8 arch/parisc/include/asm/tlb.h #define __pte_free_tlb(tlb, pte, addr) pte_free((tlb)->mm, pte) pte 77 arch/parisc/kernel/pci-dma.c static inline int map_pte_uncached(pte_t * pte, pte 91 arch/parisc/kernel/pci-dma.c if (!pte_none(*pte)) pte 94 arch/parisc/kernel/pci-dma.c set_pte(pte, __mk_pte(*paddr_ptr, PAGE_KERNEL_UNC)); pte 100 arch/parisc/kernel/pci-dma.c pte++; pte 116 arch/parisc/kernel/pci-dma.c pte_t * pte = pte_alloc_kernel(pmd, vaddr); pte 117 arch/parisc/kernel/pci-dma.c if (!pte) pte 119 arch/parisc/kernel/pci-dma.c if (map_pte_uncached(pte, orig_vaddr, end - vaddr, paddr_ptr)) pte 152 arch/parisc/kernel/pci-dma.c pte_t * pte; pte 163 arch/parisc/kernel/pci-dma.c pte = pte_offset_map(pmd, vaddr); pte 170 arch/parisc/kernel/pci-dma.c pte_t page = *pte; pte 172 arch/parisc/kernel/pci-dma.c pte_clear(&init_mm, vaddr, pte); pte 178 arch/parisc/kernel/pci-dma.c pte++; pte 18 arch/parisc/mm/fixmap.c pte_t *pte; pte 23 arch/parisc/mm/fixmap.c pte = pte_offset_kernel(pmd, vaddr); pte 24 arch/parisc/mm/fixmap.c if (pte_none(*pte)) pte 25 arch/parisc/mm/fixmap.c pte = pte_alloc_kernel(pmd, vaddr); pte 27 arch/parisc/mm/fixmap.c set_pte_at(&init_mm, vaddr, pte, __mk_pte(phys, PAGE_KERNEL_RWX)); pte 36 arch/parisc/mm/fixmap.c pte_t *pte = pte_offset_kernel(pmd, vaddr); pte 38 arch/parisc/mm/fixmap.c if (WARN_ON(pte_none(*pte))) pte 41 arch/parisc/mm/fixmap.c pte_clear(&init_mm, vaddr, pte); pte 54 arch/parisc/mm/hugetlbpage.c pte_t *pte = NULL; pte 68 arch/parisc/mm/hugetlbpage.c pte = pte_alloc_map(mm, pmd, addr); pte 70 arch/parisc/mm/hugetlbpage.c return pte; pte 79 arch/parisc/mm/hugetlbpage.c pte_t *pte = NULL; pte 89 arch/parisc/mm/hugetlbpage.c pte = pte_offset_map(pmd, addr); pte 92 arch/parisc/mm/hugetlbpage.c return pte; pte 177 arch/parisc/mm/hugetlbpage.c pte_t pte, int dirty) pte 184 arch/parisc/mm/hugetlbpage.c changed = !pte_same(*ptep, pte); pte 186 arch/parisc/mm/hugetlbpage.c __set_huge_pte_at(mm, addr, ptep, pte); pte 432 arch/parisc/mm/init.c pte_t pte; pte 454 arch/parisc/mm/init.c pte = __mk_pte(address, prot); pte 456 arch/parisc/mm/init.c pte = pte_mkhuge(pte); pte 461 arch/parisc/mm/init.c set_pte(pg_table, pte); pte 29 arch/powerpc/include/asm/book3s/32/pgalloc.h pte_t *pte) pte 31 arch/powerpc/include/asm/book3s/32/pgalloc.h *pmdp = __pmd(__pa(pte) | _PMD_PRESENT); pte 21 arch/powerpc/include/asm/book3s/32/pgtable.h static inline bool pte_user(pte_t pte) pte 23 arch/powerpc/include/asm/book3s/32/pgtable.h return pte_val(pte) & _PAGE_USER; pte 367 arch/powerpc/include/asm/book3s/32/pgtable.h #define pte_unmap(pte) kunmap_atomic(pte) pte 378 arch/powerpc/include/asm/book3s/32/pgtable.h #define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val(pte) >> 3 }) pte 382 arch/powerpc/include/asm/book3s/32/pgtable.h static inline int pte_write(pte_t pte) { return !!(pte_val(pte) & _PAGE_RW);} pte 383 arch/powerpc/include/asm/book3s/32/pgtable.h static inline int pte_read(pte_t pte) { return 1; } pte 384 arch/powerpc/include/asm/book3s/32/pgtable.h static inline int pte_dirty(pte_t pte) { return !!(pte_val(pte) & _PAGE_DIRTY); } pte 385 arch/powerpc/include/asm/book3s/32/pgtable.h static inline int pte_young(pte_t pte) { return !!(pte_val(pte) & _PAGE_ACCESSED); } pte 386 arch/powerpc/include/asm/book3s/32/pgtable.h static inline int pte_special(pte_t pte) { return !!(pte_val(pte) & _PAGE_SPECIAL); } pte 387 arch/powerpc/include/asm/book3s/32/pgtable.h static inline int pte_none(pte_t pte) { return (pte_val(pte) & ~_PTE_NONE_MASK) == 0; } pte 388 arch/powerpc/include/asm/book3s/32/pgtable.h static inline bool pte_exec(pte_t pte) { return pte_val(pte) & _PAGE_EXEC; } pte 390 arch/powerpc/include/asm/book3s/32/pgtable.h static inline int pte_present(pte_t pte) pte 392 arch/powerpc/include/asm/book3s/32/pgtable.h return pte_val(pte) & _PAGE_PRESENT; pte 395 arch/powerpc/include/asm/book3s/32/pgtable.h static inline bool pte_hw_valid(pte_t pte) pte 397 arch/powerpc/include/asm/book3s/32/pgtable.h return pte_val(pte) & _PAGE_PRESENT; pte 400 arch/powerpc/include/asm/book3s/32/pgtable.h static inline bool pte_hashpte(pte_t pte) pte 402 arch/powerpc/include/asm/book3s/32/pgtable.h return !!(pte_val(pte) & _PAGE_HASHPTE); pte 405 arch/powerpc/include/asm/book3s/32/pgtable.h static inline bool pte_ci(pte_t pte) pte 407 arch/powerpc/include/asm/book3s/32/pgtable.h return !!(pte_val(pte) & _PAGE_NO_CACHE); pte 415 arch/powerpc/include/asm/book3s/32/pgtable.h static inline bool pte_access_permitted(pte_t pte, bool write) pte 421 arch/powerpc/include/asm/book3s/32/pgtable.h if (!pte_present(pte) || !pte_user(pte) || !pte_read(pte)) pte 424 arch/powerpc/include/asm/book3s/32/pgtable.h if (write && !pte_write(pte)) pte 442 arch/powerpc/include/asm/book3s/32/pgtable.h static inline unsigned long pte_pfn(pte_t pte) pte 444 arch/powerpc/include/asm/book3s/32/pgtable.h return pte_val(pte) >> PTE_RPN_SHIFT; pte 448 arch/powerpc/include/asm/book3s/32/pgtable.h static inline pte_t pte_wrprotect(pte_t pte) pte 450 arch/powerpc/include/asm/book3s/32/pgtable.h return __pte(pte_val(pte) & ~_PAGE_RW); pte 453 arch/powerpc/include/asm/book3s/32/pgtable.h static inline pte_t pte_exprotect(pte_t pte) pte 455 arch/powerpc/include/asm/book3s/32/pgtable.h return __pte(pte_val(pte) & ~_PAGE_EXEC); pte 458 arch/powerpc/include/asm/book3s/32/pgtable.h static inline pte_t pte_mkclean(pte_t pte) pte 460 arch/powerpc/include/asm/book3s/32/pgtable.h return __pte(pte_val(pte) & ~_PAGE_DIRTY); pte 463 arch/powerpc/include/asm/book3s/32/pgtable.h static inline pte_t pte_mkold(pte_t pte) pte 465 arch/powerpc/include/asm/book3s/32/pgtable.h return __pte(pte_val(pte) & ~_PAGE_ACCESSED); pte 468 arch/powerpc/include/asm/book3s/32/pgtable.h static inline pte_t pte_mkexec(pte_t pte) pte 470 arch/powerpc/include/asm/book3s/32/pgtable.h return __pte(pte_val(pte) | _PAGE_EXEC); pte 473 arch/powerpc/include/asm/book3s/32/pgtable.h static inline pte_t pte_mkpte(pte_t pte) pte 475 arch/powerpc/include/asm/book3s/32/pgtable.h return pte; pte 478 arch/powerpc/include/asm/book3s/32/pgtable.h static inline pte_t pte_mkwrite(pte_t pte) pte 480 arch/powerpc/include/asm/book3s/32/pgtable.h return __pte(pte_val(pte) | _PAGE_RW); pte 483 arch/powerpc/include/asm/book3s/32/pgtable.h static inline pte_t pte_mkdirty(pte_t pte) pte 485 arch/powerpc/include/asm/book3s/32/pgtable.h return __pte(pte_val(pte) | _PAGE_DIRTY); pte 488 arch/powerpc/include/asm/book3s/32/pgtable.h static inline pte_t pte_mkyoung(pte_t pte) pte 490 arch/powerpc/include/asm/book3s/32/pgtable.h return __pte(pte_val(pte) | _PAGE_ACCESSED); pte 493 arch/powerpc/include/asm/book3s/32/pgtable.h static inline pte_t pte_mkspecial(pte_t pte) pte 495 arch/powerpc/include/asm/book3s/32/pgtable.h return __pte(pte_val(pte) | _PAGE_SPECIAL); pte 498 arch/powerpc/include/asm/book3s/32/pgtable.h static inline pte_t pte_mkhuge(pte_t pte) pte 500 arch/powerpc/include/asm/book3s/32/pgtable.h return pte; pte 503 arch/powerpc/include/asm/book3s/32/pgtable.h static inline pte_t pte_mkprivileged(pte_t pte) pte 505 arch/powerpc/include/asm/book3s/32/pgtable.h return __pte(pte_val(pte) & ~_PAGE_USER); pte 508 arch/powerpc/include/asm/book3s/32/pgtable.h static inline pte_t pte_mkuser(pte_t pte) pte 510 arch/powerpc/include/asm/book3s/32/pgtable.h return __pte(pte_val(pte) | _PAGE_USER); pte 513 arch/powerpc/include/asm/book3s/32/pgtable.h static inline pte_t pte_modify(pte_t pte, pgprot_t newprot) pte 515 arch/powerpc/include/asm/book3s/32/pgtable.h return __pte((pte_val(pte) & _PAGE_CHG_MASK) | pgprot_val(newprot)); pte 526 arch/powerpc/include/asm/book3s/32/pgtable.h pte_t *ptep, pte_t pte, int percpu) pte 537 arch/powerpc/include/asm/book3s/32/pgtable.h | (pte_val(pte) & ~_PAGE_HASHPTE)); pte 539 arch/powerpc/include/asm/book3s/32/pgtable.h pte_update(ptep, ~_PAGE_HASHPTE, pte_val(pte)); pte 552 arch/powerpc/include/asm/book3s/32/pgtable.h | (pte_val(pte) & ~_PAGE_HASHPTE)); pte 562 arch/powerpc/include/asm/book3s/32/pgtable.h : "r" (pte) : "memory"); pte 571 arch/powerpc/include/asm/book3s/32/pgtable.h | (pte_val(pte) & ~_PAGE_HASHPTE)); pte 79 arch/powerpc/include/asm/book3s/64/hash-64k.h static inline real_pte_t __real_pte(pte_t pte, pte_t *ptep, int offset) pte 84 arch/powerpc/include/asm/book3s/64/hash-64k.h rpte.pte = pte; pte 139 arch/powerpc/include/asm/book3s/64/hash-64k.h #define __rpte_to_pte(r) ((r).pte) pte 157 arch/powerpc/include/asm/book3s/64/hash-64k.h #define pte_pagesize_index(mm, addr, pte) \ pte 158 arch/powerpc/include/asm/book3s/64/hash-64k.h (((pte) & H_PAGE_COMBO)? MMU_PAGE_4K: MMU_PAGE_64K) pte 147 arch/powerpc/include/asm/book3s/64/hash.h pte_t *ptep, unsigned long pte, int huge); pte 211 arch/powerpc/include/asm/book3s/64/hash.h static inline int hash__pte_none(pte_t pte) pte 213 arch/powerpc/include/asm/book3s/64/hash.h return (pte_val(pte) & ~H_PTE_NONE_MASK) == 0; pte 225 arch/powerpc/include/asm/book3s/64/hash.h pte_t *ptep, pte_t pte, int percpu) pte 231 arch/powerpc/include/asm/book3s/64/hash.h *ptep = pte; pte 18 arch/powerpc/include/asm/book3s/64/hugetlb.h pte_t old_pte, pte_t pte); pte 463 arch/powerpc/include/asm/book3s/64/mmu-hash.h unsigned int hash_page_do_lazy_icache(unsigned int pp, pte_t pte, int trap); pte 489 arch/powerpc/include/asm/book3s/64/mmu-hash.h unsigned long pte); pte 153 arch/powerpc/include/asm/book3s/64/pgalloc.h pte_t *pte) pte 155 arch/powerpc/include/asm/book3s/64/pgalloc.h *pmd = __pmd(__pgtable_ptr_val(pte) | PMD_VAL_BITS); pte 335 arch/powerpc/include/asm/book3s/64/pgtable.h #define __rpte_to_pte(r) ((r).pte) pte 349 arch/powerpc/include/asm/book3s/64/pgtable.h #define pte_pagesize_index(mm, addr, pte) MMU_PAGE_4K pte 390 arch/powerpc/include/asm/book3s/64/pgtable.h static inline int __pte_write(pte_t pte) pte 392 arch/powerpc/include/asm/book3s/64/pgtable.h return !!(pte_raw(pte) & cpu_to_be64(_PAGE_WRITE)); pte 397 arch/powerpc/include/asm/book3s/64/pgtable.h static inline bool pte_savedwrite(pte_t pte) pte 406 arch/powerpc/include/asm/book3s/64/pgtable.h return !(pte_raw(pte) & cpu_to_be64(_PAGE_RWX | _PAGE_PRIVILEGED)); pte 410 arch/powerpc/include/asm/book3s/64/pgtable.h static inline bool pte_savedwrite(pte_t pte) pte 416 arch/powerpc/include/asm/book3s/64/pgtable.h static inline int pte_write(pte_t pte) pte 418 arch/powerpc/include/asm/book3s/64/pgtable.h return __pte_write(pte) || pte_savedwrite(pte); pte 421 arch/powerpc/include/asm/book3s/64/pgtable.h static inline int pte_read(pte_t pte) pte 423 arch/powerpc/include/asm/book3s/64/pgtable.h return !!(pte_raw(pte) & cpu_to_be64(_PAGE_READ)); pte 480 arch/powerpc/include/asm/book3s/64/pgtable.h static inline int pte_dirty(pte_t pte) pte 482 arch/powerpc/include/asm/book3s/64/pgtable.h return !!(pte_raw(pte) & cpu_to_be64(_PAGE_DIRTY)); pte 485 arch/powerpc/include/asm/book3s/64/pgtable.h static inline int pte_young(pte_t pte) pte 487 arch/powerpc/include/asm/book3s/64/pgtable.h return !!(pte_raw(pte) & cpu_to_be64(_PAGE_ACCESSED)); pte 490 arch/powerpc/include/asm/book3s/64/pgtable.h static inline int pte_special(pte_t pte) pte 492 arch/powerpc/include/asm/book3s/64/pgtable.h return !!(pte_raw(pte) & cpu_to_be64(_PAGE_SPECIAL)); pte 495 arch/powerpc/include/asm/book3s/64/pgtable.h static inline bool pte_exec(pte_t pte) pte 497 arch/powerpc/include/asm/book3s/64/pgtable.h return !!(pte_raw(pte) & cpu_to_be64(_PAGE_EXEC)); pte 502 arch/powerpc/include/asm/book3s/64/pgtable.h static inline bool pte_soft_dirty(pte_t pte) pte 504 arch/powerpc/include/asm/book3s/64/pgtable.h return !!(pte_raw(pte) & cpu_to_be64(_PAGE_SOFT_DIRTY)); pte 507 arch/powerpc/include/asm/book3s/64/pgtable.h static inline pte_t pte_mksoft_dirty(pte_t pte) pte 509 arch/powerpc/include/asm/book3s/64/pgtable.h return __pte_raw(pte_raw(pte) | cpu_to_be64(_PAGE_SOFT_DIRTY)); pte 512 arch/powerpc/include/asm/book3s/64/pgtable.h static inline pte_t pte_clear_soft_dirty(pte_t pte) pte 514 arch/powerpc/include/asm/book3s/64/pgtable.h return __pte_raw(pte_raw(pte) & cpu_to_be64(~_PAGE_SOFT_DIRTY)); pte 519 arch/powerpc/include/asm/book3s/64/pgtable.h static inline int pte_protnone(pte_t pte) pte 521 arch/powerpc/include/asm/book3s/64/pgtable.h return (pte_raw(pte) & cpu_to_be64(_PAGE_PRESENT | _PAGE_PTE | _PAGE_RWX)) == pte 526 arch/powerpc/include/asm/book3s/64/pgtable.h static inline pte_t pte_mk_savedwrite(pte_t pte) pte 533 arch/powerpc/include/asm/book3s/64/pgtable.h VM_BUG_ON((pte_raw(pte) & cpu_to_be64(_PAGE_PRESENT | _PAGE_RWX | _PAGE_PRIVILEGED)) != pte 535 arch/powerpc/include/asm/book3s/64/pgtable.h return __pte_raw(pte_raw(pte) & cpu_to_be64(~_PAGE_PRIVILEGED)); pte 539 arch/powerpc/include/asm/book3s/64/pgtable.h static inline pte_t pte_clear_savedwrite(pte_t pte) pte 544 arch/powerpc/include/asm/book3s/64/pgtable.h VM_BUG_ON(!pte_protnone(pte)); pte 545 arch/powerpc/include/asm/book3s/64/pgtable.h return __pte_raw(pte_raw(pte) | cpu_to_be64(_PAGE_PRIVILEGED)); pte 549 arch/powerpc/include/asm/book3s/64/pgtable.h static inline pte_t pte_clear_savedwrite(pte_t pte) pte 552 arch/powerpc/include/asm/book3s/64/pgtable.h return __pte_raw(pte_raw(pte) & cpu_to_be64(~_PAGE_WRITE)); pte 556 arch/powerpc/include/asm/book3s/64/pgtable.h static inline int pte_present(pte_t pte) pte 564 arch/powerpc/include/asm/book3s/64/pgtable.h return !!(pte_raw(pte) & cpu_to_be64(_PAGE_PRESENT | _PAGE_INVALID)); pte 567 arch/powerpc/include/asm/book3s/64/pgtable.h static inline bool pte_hw_valid(pte_t pte) pte 569 arch/powerpc/include/asm/book3s/64/pgtable.h return !!(pte_raw(pte) & cpu_to_be64(_PAGE_PRESENT)); pte 573 arch/powerpc/include/asm/book3s/64/pgtable.h extern bool arch_pte_access_permitted(u64 pte, bool write, bool execute); pte 575 arch/powerpc/include/asm/book3s/64/pgtable.h static inline bool arch_pte_access_permitted(u64 pte, bool write, bool execute) pte 581 arch/powerpc/include/asm/book3s/64/pgtable.h static inline bool pte_user(pte_t pte) pte 583 arch/powerpc/include/asm/book3s/64/pgtable.h return !(pte_raw(pte) & cpu_to_be64(_PAGE_PRIVILEGED)); pte 587 arch/powerpc/include/asm/book3s/64/pgtable.h static inline bool pte_access_permitted(pte_t pte, bool write) pte 593 arch/powerpc/include/asm/book3s/64/pgtable.h if (!pte_present(pte) || !pte_user(pte) || !pte_read(pte)) pte 596 arch/powerpc/include/asm/book3s/64/pgtable.h if (write && !pte_write(pte)) pte 599 arch/powerpc/include/asm/book3s/64/pgtable.h return arch_pte_access_permitted(pte_val(pte), write, 0); pte 617 arch/powerpc/include/asm/book3s/64/pgtable.h static inline unsigned long pte_pfn(pte_t pte) pte 619 arch/powerpc/include/asm/book3s/64/pgtable.h return (pte_val(pte) & PTE_RPN_MASK) >> PAGE_SHIFT; pte 623 arch/powerpc/include/asm/book3s/64/pgtable.h static inline pte_t pte_wrprotect(pte_t pte) pte 625 arch/powerpc/include/asm/book3s/64/pgtable.h if (unlikely(pte_savedwrite(pte))) pte 626 arch/powerpc/include/asm/book3s/64/pgtable.h return pte_clear_savedwrite(pte); pte 627 arch/powerpc/include/asm/book3s/64/pgtable.h return __pte_raw(pte_raw(pte) & cpu_to_be64(~_PAGE_WRITE)); pte 630 arch/powerpc/include/asm/book3s/64/pgtable.h static inline pte_t pte_exprotect(pte_t pte) pte 632 arch/powerpc/include/asm/book3s/64/pgtable.h return __pte_raw(pte_raw(pte) & cpu_to_be64(~_PAGE_EXEC)); pte 635 arch/powerpc/include/asm/book3s/64/pgtable.h static inline pte_t pte_mkclean(pte_t pte) pte 637 arch/powerpc/include/asm/book3s/64/pgtable.h return __pte_raw(pte_raw(pte) & cpu_to_be64(~_PAGE_DIRTY)); pte 640 arch/powerpc/include/asm/book3s/64/pgtable.h static inline pte_t pte_mkold(pte_t pte) pte 642 arch/powerpc/include/asm/book3s/64/pgtable.h return __pte_raw(pte_raw(pte) & cpu_to_be64(~_PAGE_ACCESSED)); pte 645 arch/powerpc/include/asm/book3s/64/pgtable.h static inline pte_t pte_mkexec(pte_t pte) pte 647 arch/powerpc/include/asm/book3s/64/pgtable.h return __pte_raw(pte_raw(pte) | cpu_to_be64(_PAGE_EXEC)); pte 650 arch/powerpc/include/asm/book3s/64/pgtable.h static inline pte_t pte_mkpte(pte_t pte) pte 652 arch/powerpc/include/asm/book3s/64/pgtable.h return __pte_raw(pte_raw(pte) | cpu_to_be64(_PAGE_PTE)); pte 655 arch/powerpc/include/asm/book3s/64/pgtable.h static inline pte_t pte_mkwrite(pte_t pte) pte 660 arch/powerpc/include/asm/book3s/64/pgtable.h return __pte_raw(pte_raw(pte) | cpu_to_be64(_PAGE_RW)); pte 663 arch/powerpc/include/asm/book3s/64/pgtable.h static inline pte_t pte_mkdirty(pte_t pte) pte 665 arch/powerpc/include/asm/book3s/64/pgtable.h return __pte_raw(pte_raw(pte) | cpu_to_be64(_PAGE_DIRTY | _PAGE_SOFT_DIRTY)); pte 668 arch/powerpc/include/asm/book3s/64/pgtable.h static inline pte_t pte_mkyoung(pte_t pte) pte 670 arch/powerpc/include/asm/book3s/64/pgtable.h return __pte_raw(pte_raw(pte) | cpu_to_be64(_PAGE_ACCESSED)); pte 673 arch/powerpc/include/asm/book3s/64/pgtable.h static inline pte_t pte_mkspecial(pte_t pte) pte 675 arch/powerpc/include/asm/book3s/64/pgtable.h return __pte_raw(pte_raw(pte) | cpu_to_be64(_PAGE_SPECIAL)); pte 678 arch/powerpc/include/asm/book3s/64/pgtable.h static inline pte_t pte_mkhuge(pte_t pte) pte 680 arch/powerpc/include/asm/book3s/64/pgtable.h return pte; pte 683 arch/powerpc/include/asm/book3s/64/pgtable.h static inline pte_t pte_mkdevmap(pte_t pte) pte 685 arch/powerpc/include/asm/book3s/64/pgtable.h return __pte_raw(pte_raw(pte) | cpu_to_be64(_PAGE_SPECIAL | _PAGE_DEVMAP)); pte 688 arch/powerpc/include/asm/book3s/64/pgtable.h static inline pte_t pte_mkprivileged(pte_t pte) pte 690 arch/powerpc/include/asm/book3s/64/pgtable.h return __pte_raw(pte_raw(pte) | cpu_to_be64(_PAGE_PRIVILEGED)); pte 693 arch/powerpc/include/asm/book3s/64/pgtable.h static inline pte_t pte_mkuser(pte_t pte) pte 695 arch/powerpc/include/asm/book3s/64/pgtable.h return __pte_raw(pte_raw(pte) & cpu_to_be64(~_PAGE_PRIVILEGED)); pte 704 arch/powerpc/include/asm/book3s/64/pgtable.h static inline int pte_devmap(pte_t pte) pte 708 arch/powerpc/include/asm/book3s/64/pgtable.h return (pte_raw(pte) & mask) == mask; pte 711 arch/powerpc/include/asm/book3s/64/pgtable.h static inline pte_t pte_modify(pte_t pte, pgprot_t newprot) pte 714 arch/powerpc/include/asm/book3s/64/pgtable.h return __pte_raw((pte_raw(pte) & cpu_to_be64(_PAGE_CHG_MASK)) | pte 742 arch/powerpc/include/asm/book3s/64/pgtable.h #define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val((pte)) & ~_PAGE_PTE }) pte 754 arch/powerpc/include/asm/book3s/64/pgtable.h static inline pte_t pte_swp_mksoft_dirty(pte_t pte) pte 756 arch/powerpc/include/asm/book3s/64/pgtable.h return __pte_raw(pte_raw(pte) | cpu_to_be64(_PAGE_SWP_SOFT_DIRTY)); pte 759 arch/powerpc/include/asm/book3s/64/pgtable.h static inline bool pte_swp_soft_dirty(pte_t pte) pte 761 arch/powerpc/include/asm/book3s/64/pgtable.h return !!(pte_raw(pte) & cpu_to_be64(_PAGE_SWP_SOFT_DIRTY)); pte 764 arch/powerpc/include/asm/book3s/64/pgtable.h static inline pte_t pte_swp_clear_soft_dirty(pte_t pte) pte 766 arch/powerpc/include/asm/book3s/64/pgtable.h return __pte_raw(pte_raw(pte) & cpu_to_be64(~_PAGE_SWP_SOFT_DIRTY)); pte 808 arch/powerpc/include/asm/book3s/64/pgtable.h static inline int pte_none(pte_t pte) pte 811 arch/powerpc/include/asm/book3s/64/pgtable.h return radix__pte_none(pte); pte 812 arch/powerpc/include/asm/book3s/64/pgtable.h return hash__pte_none(pte); pte 816 arch/powerpc/include/asm/book3s/64/pgtable.h pte_t *ptep, pte_t pte, int percpu) pte 819 arch/powerpc/include/asm/book3s/64/pgtable.h return radix__set_pte_at(mm, addr, ptep, pte, percpu); pte 820 arch/powerpc/include/asm/book3s/64/pgtable.h return hash__set_pte_at(mm, addr, ptep, pte, percpu); pte 853 arch/powerpc/include/asm/book3s/64/pgtable.h static inline bool pte_ci(pte_t pte) pte 855 arch/powerpc/include/asm/book3s/64/pgtable.h __be64 pte_v = pte_raw(pte); pte 933 arch/powerpc/include/asm/book3s/64/pgtable.h static inline pud_t pte_pud(pte_t pte) pte 935 arch/powerpc/include/asm/book3s/64/pgtable.h return __pud_raw(pte_raw(pte)); pte 974 arch/powerpc/include/asm/book3s/64/pgtable.h static inline pgd_t pte_pgd(pte_t pte) pte 976 arch/powerpc/include/asm/book3s/64/pgtable.h return __pgd_raw(pte_raw(pte)); pte 1022 arch/powerpc/include/asm/book3s/64/pgtable.h static inline void pte_unmap(pte_t *pte) { } pte 1073 arch/powerpc/include/asm/book3s/64/pgtable.h static inline pmd_t pte_pmd(pte_t pte) pte 1075 arch/powerpc/include/asm/book3s/64/pgtable.h return __pmd_raw(pte_raw(pte)); pte 130 arch/powerpc/include/asm/book3s/64/radix.h pte_t old_pte, pte_t pte); pte 185 arch/powerpc/include/asm/book3s/64/radix.h static inline int radix__pte_none(pte_t pte) pte 187 arch/powerpc/include/asm/book3s/64/radix.h return (pte_val(pte) & ~RADIX_PTE_NONE_MASK) == 0; pte 191 arch/powerpc/include/asm/book3s/64/radix.h pte_t *ptep, pte_t pte, int percpu) pte 193 arch/powerpc/include/asm/book3s/64/radix.h *ptep = pte; pte 18 arch/powerpc/include/asm/book3s/64/tlbflush-hash.h real_pte_t pte[PPC64_TLB_BATCH_NR]; pte 56 arch/powerpc/include/asm/book3s/64/tlbflush-hash.h extern void flush_hash_page(unsigned long vpn, real_pte_t pte, int psize, pte 17 arch/powerpc/include/asm/book3s/pgtable.h pte_t pte); pte 61 arch/powerpc/include/asm/hugetlb.h pte_t pte, int dirty); pte 57 arch/powerpc/include/asm/kvm_book3s.h struct kvmppc_pte pte; pte 152 arch/powerpc/include/asm/kvm_book3s.h extern int kvmppc_mmu_map_page(struct kvm_vcpu *vcpu, struct kvmppc_pte *pte, pte 154 arch/powerpc/include/asm/kvm_book3s.h extern void kvmppc_mmu_unmap_page(struct kvm_vcpu *vcpu, struct kvmppc_pte *pte); pte 166 arch/powerpc/include/asm/kvm_book3s.h extern void kvmppc_mmu_hpte_cache_map(struct kvm_vcpu *vcpu, struct hpte_cache *pte); pte 168 arch/powerpc/include/asm/kvm_book3s.h extern void kvmppc_mmu_hpte_cache_free(struct hpte_cache *pte); pte 171 arch/powerpc/include/asm/kvm_book3s.h extern void kvmppc_mmu_invalidate_pte(struct kvm_vcpu *vcpu, struct hpte_cache *pte); pte 197 arch/powerpc/include/asm/kvm_book3s.h extern void kvmppc_unmap_pte(struct kvm *kvm, pte_t *pte, unsigned long gpa, pte 623 arch/powerpc/include/asm/kvm_book3s_64.h extern int kvmppc_create_pte(struct kvm *kvm, pgd_t *pgtable, pte_t pte, pte 64 arch/powerpc/include/asm/kvm_host.h extern int kvm_set_spte_hva(struct kvm *kvm, unsigned long hva, pte_t pte); pte 403 arch/powerpc/include/asm/kvm_host.h struct kvmppc_pte *pte, bool data, bool iswrite); pte 120 arch/powerpc/include/asm/kvm_ppc.h struct kvmppc_pte *pte); pte 293 arch/powerpc/include/asm/kvm_ppc.h void (*set_spte_hva)(struct kvm *kvm, unsigned long hva, pte_t pte); pte 18 arch/powerpc/include/asm/nohash/32/pgalloc.h pte_t *pte) pte 21 arch/powerpc/include/asm/nohash/32/pgalloc.h *pmdp = __pmd((unsigned long)pte | _PMD_PRESENT); pte 23 arch/powerpc/include/asm/nohash/32/pgalloc.h *pmdp = __pmd(__pa(pte) | _PMD_PRESENT); pte 167 arch/powerpc/include/asm/nohash/32/pgtable.h static inline pte_t pte_mkwrite(pte_t pte) pte 169 arch/powerpc/include/asm/nohash/32/pgtable.h return __pte(pte_val(pte) | _PAGE_RW); pte 173 arch/powerpc/include/asm/nohash/32/pgtable.h static inline pte_t pte_mkdirty(pte_t pte) pte 175 arch/powerpc/include/asm/nohash/32/pgtable.h return __pte(pte_val(pte) | _PAGE_DIRTY); pte 178 arch/powerpc/include/asm/nohash/32/pgtable.h static inline pte_t pte_mkyoung(pte_t pte) pte 180 arch/powerpc/include/asm/nohash/32/pgtable.h return __pte(pte_val(pte) | _PAGE_ACCESSED); pte 184 arch/powerpc/include/asm/nohash/32/pgtable.h static inline pte_t pte_wrprotect(pte_t pte) pte 186 arch/powerpc/include/asm/nohash/32/pgtable.h return __pte(pte_val(pte) & ~_PAGE_RW); pte 190 arch/powerpc/include/asm/nohash/32/pgtable.h static inline pte_t pte_mkexec(pte_t pte) pte 192 arch/powerpc/include/asm/nohash/32/pgtable.h return __pte(pte_val(pte) | _PAGE_EXEC); pte 243 arch/powerpc/include/asm/nohash/32/pgtable.h p->pte = p->pte1 = p->pte2 = p->pte3 = new; pte 330 arch/powerpc/include/asm/nohash/32/pgtable.h static inline int pte_young(pte_t pte) pte 332 arch/powerpc/include/asm/nohash/32/pgtable.h return pte_val(pte) & _PAGE_ACCESSED; pte 373 arch/powerpc/include/asm/nohash/32/pgtable.h #define pte_unmap(pte) kunmap_atomic(pte) pte 384 arch/powerpc/include/asm/nohash/32/pgtable.h #define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val(pte) >> 3 }) pte 89 arch/powerpc/include/asm/nohash/32/pte-40x.h static inline pte_t pte_wrprotect(pte_t pte) pte 91 arch/powerpc/include/asm/nohash/32/pte-40x.h return __pte(pte_val(pte) & ~(_PAGE_RW | _PAGE_HWWRITE)); pte 96 arch/powerpc/include/asm/nohash/32/pte-40x.h static inline pte_t pte_mkclean(pte_t pte) pte 98 arch/powerpc/include/asm/nohash/32/pte-40x.h return __pte(pte_val(pte) & ~(_PAGE_DIRTY | _PAGE_HWWRITE)); pte 87 arch/powerpc/include/asm/nohash/32/pte-8xx.h static inline pte_t pte_wrprotect(pte_t pte) pte 89 arch/powerpc/include/asm/nohash/32/pte-8xx.h return __pte(pte_val(pte) | _PAGE_RO); pte 94 arch/powerpc/include/asm/nohash/32/pte-8xx.h static inline int pte_write(pte_t pte) pte 96 arch/powerpc/include/asm/nohash/32/pte-8xx.h return !(pte_val(pte) & _PAGE_RO); pte 101 arch/powerpc/include/asm/nohash/32/pte-8xx.h static inline pte_t pte_mkwrite(pte_t pte) pte 103 arch/powerpc/include/asm/nohash/32/pte-8xx.h return __pte(pte_val(pte) & ~_PAGE_RO); pte 108 arch/powerpc/include/asm/nohash/32/pte-8xx.h static inline bool pte_user(pte_t pte) pte 110 arch/powerpc/include/asm/nohash/32/pte-8xx.h return !(pte_val(pte) & _PAGE_SH); pte 115 arch/powerpc/include/asm/nohash/32/pte-8xx.h static inline pte_t pte_mkprivileged(pte_t pte) pte 117 arch/powerpc/include/asm/nohash/32/pte-8xx.h return __pte(pte_val(pte) | _PAGE_SH); pte 122 arch/powerpc/include/asm/nohash/32/pte-8xx.h static inline pte_t pte_mkuser(pte_t pte) pte 124 arch/powerpc/include/asm/nohash/32/pte-8xx.h return __pte(pte_val(pte) & ~_PAGE_SH); pte 129 arch/powerpc/include/asm/nohash/32/pte-8xx.h static inline pte_t pte_mkhuge(pte_t pte) pte 131 arch/powerpc/include/asm/nohash/32/pte-8xx.h return __pte(pte_val(pte) | _PAGE_SPS); pte 37 arch/powerpc/include/asm/nohash/64/pgalloc.h pte_t *pte) pte 39 arch/powerpc/include/asm/nohash/64/pgalloc.h pmd_set(pmd, (unsigned long)pte); pte 73 arch/powerpc/include/asm/nohash/64/pgtable-4k.h static inline pgd_t pte_pgd(pte_t pte) pte 75 arch/powerpc/include/asm/nohash/64/pgtable-4k.h return __pgd(pte_val(pte)); pte 100 arch/powerpc/include/asm/nohash/64/pgtable.h static inline pte_t pte_mkwrite(pte_t pte) pte 102 arch/powerpc/include/asm/nohash/64/pgtable.h return __pte(pte_val(pte) | _PAGE_RW); pte 105 arch/powerpc/include/asm/nohash/64/pgtable.h static inline pte_t pte_mkdirty(pte_t pte) pte 107 arch/powerpc/include/asm/nohash/64/pgtable.h return __pte(pte_val(pte) | _PAGE_DIRTY); pte 110 arch/powerpc/include/asm/nohash/64/pgtable.h static inline pte_t pte_mkyoung(pte_t pte) pte 112 arch/powerpc/include/asm/nohash/64/pgtable.h return __pte(pte_val(pte) | _PAGE_ACCESSED); pte 115 arch/powerpc/include/asm/nohash/64/pgtable.h static inline pte_t pte_wrprotect(pte_t pte) pte 117 arch/powerpc/include/asm/nohash/64/pgtable.h return __pte(pte_val(pte) & ~_PAGE_RW); pte 120 arch/powerpc/include/asm/nohash/64/pgtable.h static inline pte_t pte_mkexec(pte_t pte) pte 122 arch/powerpc/include/asm/nohash/64/pgtable.h return __pte(pte_val(pte) | _PAGE_EXEC); pte 173 arch/powerpc/include/asm/nohash/64/pgtable.h static inline pud_t pte_pud(pte_t pte) pte 175 arch/powerpc/include/asm/nohash/64/pgtable.h return __pud(pte_val(pte)); pte 201 arch/powerpc/include/asm/nohash/64/pgtable.h static inline void pte_unmap(pte_t *pte) { } pte 237 arch/powerpc/include/asm/nohash/64/pgtable.h static inline int pte_young(pte_t pte) pte 239 arch/powerpc/include/asm/nohash/64/pgtable.h return pte_val(pte) & _PAGE_ACCESSED; pte 355 arch/powerpc/include/asm/nohash/64/pgtable.h #define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val((pte)) }) pte 43 arch/powerpc/include/asm/nohash/pgtable.h static inline int pte_write(pte_t pte) pte 45 arch/powerpc/include/asm/nohash/pgtable.h return pte_val(pte) & _PAGE_RW; pte 48 arch/powerpc/include/asm/nohash/pgtable.h static inline int pte_read(pte_t pte) { return 1; } pte 49 arch/powerpc/include/asm/nohash/pgtable.h static inline int pte_dirty(pte_t pte) { return pte_val(pte) & _PAGE_DIRTY; } pte 50 arch/powerpc/include/asm/nohash/pgtable.h static inline int pte_special(pte_t pte) { return pte_val(pte) & _PAGE_SPECIAL; } pte 51 arch/powerpc/include/asm/nohash/pgtable.h static inline int pte_none(pte_t pte) { return (pte_val(pte) & ~_PTE_NONE_MASK) == 0; } pte 52 arch/powerpc/include/asm/nohash/pgtable.h static inline bool pte_hashpte(pte_t pte) { return false; } pte 53 arch/powerpc/include/asm/nohash/pgtable.h static inline bool pte_ci(pte_t pte) { return pte_val(pte) & _PAGE_NO_CACHE; } pte 54 arch/powerpc/include/asm/nohash/pgtable.h static inline bool pte_exec(pte_t pte) { return pte_val(pte) & _PAGE_EXEC; } pte 62 arch/powerpc/include/asm/nohash/pgtable.h static inline int pte_protnone(pte_t pte) pte 64 arch/powerpc/include/asm/nohash/pgtable.h return pte_present(pte) && !pte_user(pte); pte 73 arch/powerpc/include/asm/nohash/pgtable.h static inline int pte_present(pte_t pte) pte 75 arch/powerpc/include/asm/nohash/pgtable.h return pte_val(pte) & _PAGE_PRESENT; pte 78 arch/powerpc/include/asm/nohash/pgtable.h static inline bool pte_hw_valid(pte_t pte) pte 80 arch/powerpc/include/asm/nohash/pgtable.h return pte_val(pte) & _PAGE_PRESENT; pte 89 arch/powerpc/include/asm/nohash/pgtable.h static inline bool pte_user(pte_t pte) pte 91 arch/powerpc/include/asm/nohash/pgtable.h return (pte_val(pte) & _PAGE_USER) == _PAGE_USER; pte 100 arch/powerpc/include/asm/nohash/pgtable.h static inline bool pte_access_permitted(pte_t pte, bool write) pte 106 arch/powerpc/include/asm/nohash/pgtable.h if (!pte_present(pte) || !pte_user(pte) || !pte_read(pte)) pte 109 arch/powerpc/include/asm/nohash/pgtable.h if (write && !pte_write(pte)) pte 124 arch/powerpc/include/asm/nohash/pgtable.h static inline unsigned long pte_pfn(pte_t pte) { pte 125 arch/powerpc/include/asm/nohash/pgtable.h return pte_val(pte) >> PTE_RPN_SHIFT; } pte 128 arch/powerpc/include/asm/nohash/pgtable.h static inline pte_t pte_exprotect(pte_t pte) pte 130 arch/powerpc/include/asm/nohash/pgtable.h return __pte(pte_val(pte) & ~_PAGE_EXEC); pte 134 arch/powerpc/include/asm/nohash/pgtable.h static inline pte_t pte_mkclean(pte_t pte) pte 136 arch/powerpc/include/asm/nohash/pgtable.h return __pte(pte_val(pte) & ~_PAGE_DIRTY); pte 140 arch/powerpc/include/asm/nohash/pgtable.h static inline pte_t pte_mkold(pte_t pte) pte 142 arch/powerpc/include/asm/nohash/pgtable.h return __pte(pte_val(pte) & ~_PAGE_ACCESSED); pte 145 arch/powerpc/include/asm/nohash/pgtable.h static inline pte_t pte_mkpte(pte_t pte) pte 147 arch/powerpc/include/asm/nohash/pgtable.h return pte; pte 150 arch/powerpc/include/asm/nohash/pgtable.h static inline pte_t pte_mkspecial(pte_t pte) pte 152 arch/powerpc/include/asm/nohash/pgtable.h return __pte(pte_val(pte) | _PAGE_SPECIAL); pte 156 arch/powerpc/include/asm/nohash/pgtable.h static inline pte_t pte_mkhuge(pte_t pte) pte 158 arch/powerpc/include/asm/nohash/pgtable.h return __pte(pte_val(pte)); pte 163 arch/powerpc/include/asm/nohash/pgtable.h static inline pte_t pte_mkprivileged(pte_t pte) pte 165 arch/powerpc/include/asm/nohash/pgtable.h return __pte(pte_val(pte) & ~_PAGE_USER); pte 170 arch/powerpc/include/asm/nohash/pgtable.h static inline pte_t pte_mkuser(pte_t pte) pte 172 arch/powerpc/include/asm/nohash/pgtable.h return __pte(pte_val(pte) | _PAGE_USER); pte 176 arch/powerpc/include/asm/nohash/pgtable.h static inline pte_t pte_modify(pte_t pte, pgprot_t newprot) pte 178 arch/powerpc/include/asm/nohash/pgtable.h return __pte((pte_val(pte) & _PAGE_CHG_MASK) | pgprot_val(newprot)); pte 185 arch/powerpc/include/asm/nohash/pgtable.h pte_t pte); pte 193 arch/powerpc/include/asm/nohash/pgtable.h pte_t *ptep, pte_t pte, int percpu) pte 206 arch/powerpc/include/asm/nohash/pgtable.h : "r" (pte) : "memory"); pte 213 arch/powerpc/include/asm/nohash/pgtable.h ptep->pte = ptep->pte1 = ptep->pte2 = ptep->pte3 = pte_val(pte); pte 215 arch/powerpc/include/asm/nohash/pgtable.h *ptep = pte; pte 103 arch/powerpc/include/asm/nohash/pte-book3e.h static inline pte_t pte_mkprivileged(pte_t pte) pte 105 arch/powerpc/include/asm/nohash/pte-book3e.h return __pte((pte_val(pte) & ~_PAGE_USER) | _PAGE_PRIVILEGED); pte 110 arch/powerpc/include/asm/nohash/pte-book3e.h static inline pte_t pte_mkuser(pte_t pte) pte 112 arch/powerpc/include/asm/nohash/pte-book3e.h return __pte((pte_val(pte) & ~_PAGE_PRIVILEGED) | _PAGE_USER); pte 38 arch/powerpc/include/asm/pgalloc.h static inline void pte_free_kernel(struct mm_struct *mm, pte_t *pte) pte 40 arch/powerpc/include/asm/pgalloc.h pte_fragment_free((unsigned long *)pte, 1); pte 8 arch/powerpc/include/asm/pgtable-be-types.h typedef struct { __be64 pte; } pte_t; pte 13 arch/powerpc/include/asm/pgtable-be-types.h return be64_to_cpu(x.pte); pte 18 arch/powerpc/include/asm/pgtable-be-types.h return x.pte; pte 76 arch/powerpc/include/asm/pgtable-be-types.h typedef struct { pte_t pte; unsigned long hidx; } real_pte_t; pte 78 arch/powerpc/include/asm/pgtable-be-types.h typedef struct { pte_t pte; } real_pte_t; pte 7 arch/powerpc/include/asm/pgtable-types.h typedef struct { pte_basic_t pte, pte1, pte2, pte3; } pte_t; pte 9 arch/powerpc/include/asm/pgtable-types.h typedef struct { pte_basic_t pte; } pte_t; pte 14 arch/powerpc/include/asm/pgtable-types.h return x.pte; pte 53 arch/powerpc/include/asm/pgtable-types.h typedef struct { pte_t pte; unsigned long hidx; } real_pte_t; pte 55 arch/powerpc/include/asm/pgtable-types.h typedef struct { pte_t pte; } real_pte_t; pte 52 arch/powerpc/include/asm/pgtable.h static inline pgprot_t pte_pgprot(pte_t pte) pte 56 arch/powerpc/include/asm/pgtable.h pte_flags = pte_val(pte) & ~PTE_RPN_MASK; pte 13 arch/powerpc/include/asm/pte-walk.h pte_t *pte; pte 16 arch/powerpc/include/asm/pte-walk.h pte = __find_linux_pte(pgdir, ea, is_thp, hshift); pte 26 arch/powerpc/include/asm/pte-walk.h return pte; pte 41 arch/powerpc/include/asm/pte-walk.h pte_t *pte; pte 46 arch/powerpc/include/asm/pte-walk.h pte = __find_linux_pte(pgdir, ea, is_thp, hshift); pte 56 arch/powerpc/include/asm/pte-walk.h return pte; pte 453 arch/powerpc/kvm/book3s.c enum xlate_readwrite xlrw, struct kvmppc_pte *pte) pte 461 arch/powerpc/kvm/book3s.c r = vcpu->arch.mmu.xlate(vcpu, eaddr, pte, data, iswrite); pte 463 arch/powerpc/kvm/book3s.c pte->eaddr = eaddr; pte 464 arch/powerpc/kvm/book3s.c pte->raddr = eaddr & KVM_PAM; pte 465 arch/powerpc/kvm/book3s.c pte->vpage = VSID_REAL | eaddr >> 12; pte 466 arch/powerpc/kvm/book3s.c pte->may_read = true; pte 467 arch/powerpc/kvm/book3s.c pte->may_write = true; pte 468 arch/powerpc/kvm/book3s.c pte->may_execute = true; pte 475 arch/powerpc/kvm/book3s.c pte->raddr &= ~SPLIT_HACK_MASK; pte 885 arch/powerpc/kvm/book3s.c int kvm_set_spte_hva(struct kvm *kvm, unsigned long hva, pte_t pte) pte 887 arch/powerpc/kvm/book3s.c kvm->arch.kvm_ops->set_spte_hva(kvm, hva, pte); pte 17 arch/powerpc/kvm/book3s.h extern void kvm_set_spte_hva_hv(struct kvm *kvm, unsigned long hva, pte_t pte); pte 70 arch/powerpc/kvm/book3s_32_mmu.c struct kvmppc_pte *pte, bool data, pte 84 arch/powerpc/kvm/book3s_32_mmu.c struct kvmppc_pte pte; pte 86 arch/powerpc/kvm/book3s_32_mmu.c if (!kvmppc_mmu_book3s_32_xlate_bat(vcpu, eaddr, &pte, data, false)) pte 87 arch/powerpc/kvm/book3s_32_mmu.c return pte.vpage; pte 133 arch/powerpc/kvm/book3s_32_mmu.c struct kvmppc_pte *pte, bool data, pte 165 arch/powerpc/kvm/book3s_32_mmu.c pte->vpage = (((u64)eaddr >> 12) & 0xffff) | vsid; pte 167 arch/powerpc/kvm/book3s_32_mmu.c pte->raddr = bat->brpn | (eaddr & ~bat->bepi_mask); pte 168 arch/powerpc/kvm/book3s_32_mmu.c pte->may_read = bat->pp; pte 169 arch/powerpc/kvm/book3s_32_mmu.c pte->may_write = bat->pp > 1; pte 170 arch/powerpc/kvm/book3s_32_mmu.c pte->may_execute = true; pte 171 arch/powerpc/kvm/book3s_32_mmu.c if (!pte->may_read) { pte 175 arch/powerpc/kvm/book3s_32_mmu.c if (iswrite && !pte->may_write) { pte 188 arch/powerpc/kvm/book3s_32_mmu.c struct kvmppc_pte *pte, bool data, pte 204 arch/powerpc/kvm/book3s_32_mmu.c pte->vpage = kvmppc_mmu_book3s_32_ea_to_vp(vcpu, eaddr, data); pte 226 arch/powerpc/kvm/book3s_32_mmu.c pte->raddr = (pte1 & ~(0xFFFULL)) | (eaddr & 0xFFF); pte 233 arch/powerpc/kvm/book3s_32_mmu.c pte->may_write = false; pte 234 arch/powerpc/kvm/book3s_32_mmu.c pte->may_read = false; pte 235 arch/powerpc/kvm/book3s_32_mmu.c pte->may_execute = true; pte 241 arch/powerpc/kvm/book3s_32_mmu.c pte->may_write = true; pte 246 arch/powerpc/kvm/book3s_32_mmu.c pte->may_read = true; pte 267 arch/powerpc/kvm/book3s_32_mmu.c if (pte->may_read && !(pte_r & PTEG_FLAG_ACCESSED)) { pte 271 arch/powerpc/kvm/book3s_32_mmu.c if (iswrite && pte->may_write && !(pte_r & PTEG_FLAG_DIRTY)) { pte 275 arch/powerpc/kvm/book3s_32_mmu.c if (!pte->may_read || (iswrite && !pte->may_write)) pte 296 arch/powerpc/kvm/book3s_32_mmu.c struct kvmppc_pte *pte, bool data, pte 302 arch/powerpc/kvm/book3s_32_mmu.c pte->eaddr = eaddr; pte 303 arch/powerpc/kvm/book3s_32_mmu.c pte->page_size = MMU_PAGE_4K; pte 309 arch/powerpc/kvm/book3s_32_mmu.c pte->vpage = kvmppc_mmu_book3s_32_ea_to_vp(vcpu, eaddr, data); pte 310 arch/powerpc/kvm/book3s_32_mmu.c pte->raddr = vcpu->arch.magic_page_pa | (pte->raddr & 0xfff); pte 311 arch/powerpc/kvm/book3s_32_mmu.c pte->raddr &= KVM_PAM; pte 312 arch/powerpc/kvm/book3s_32_mmu.c pte->may_execute = true; pte 313 arch/powerpc/kvm/book3s_32_mmu.c pte->may_read = true; pte 314 arch/powerpc/kvm/book3s_32_mmu.c pte->may_write = true; pte 319 arch/powerpc/kvm/book3s_32_mmu.c r = kvmppc_mmu_book3s_32_xlate_bat(vcpu, eaddr, pte, data, iswrite); pte 321 arch/powerpc/kvm/book3s_32_mmu.c r = kvmppc_mmu_book3s_32_xlate_pte(vcpu, eaddr, pte, pte 324 arch/powerpc/kvm/book3s_32_mmu.c r = kvmppc_mmu_book3s_32_xlate_pte(vcpu, eaddr, pte, pte 49 arch/powerpc/kvm/book3s_32_mmu_host.c void kvmppc_mmu_invalidate_pte(struct kvm_vcpu *vcpu, struct hpte_cache *pte) pte 54 arch/powerpc/kvm/book3s_32_mmu_host.c pteg = (u32*)pte->slot; pte 59 arch/powerpc/kvm/book3s_32_mmu_host.c asm volatile ("tlbie %0" : : "r" (pte->pte.eaddr) : "memory"); pte 143 arch/powerpc/kvm/book3s_32_mmu_host.c struct hpte_cache *pte; pte 233 arch/powerpc/kvm/book3s_32_mmu_host.c pte = kvmppc_mmu_hpte_cache_next(vcpu); pte 234 arch/powerpc/kvm/book3s_32_mmu_host.c if (!pte) { pte 246 arch/powerpc/kvm/book3s_32_mmu_host.c pte->slot = (ulong)&pteg[rr]; pte 247 arch/powerpc/kvm/book3s_32_mmu_host.c pte->host_vpn = vpn; pte 248 arch/powerpc/kvm/book3s_32_mmu_host.c pte->pte = *orig_pte; pte 249 arch/powerpc/kvm/book3s_32_mmu_host.c pte->pfn = hpaddr >> PAGE_SHIFT; pte 251 arch/powerpc/kvm/book3s_32_mmu_host.c kvmppc_mmu_hpte_cache_map(vcpu, pte); pte 258 arch/powerpc/kvm/book3s_32_mmu_host.c void kvmppc_mmu_unmap_page(struct kvm_vcpu *vcpu, struct kvmppc_pte *pte) pte 260 arch/powerpc/kvm/book3s_32_mmu_host.c kvmppc_mmu_pte_vflush(vcpu, pte->vpage, 0xfffffffffULL); pte 23 arch/powerpc/kvm/book3s_64_mmu_host.c void kvmppc_mmu_invalidate_pte(struct kvm_vcpu *vcpu, struct hpte_cache *pte) pte 25 arch/powerpc/kvm/book3s_64_mmu_host.c mmu_hash_ops.hpte_invalidate(pte->slot, pte->host_vpn, pte 26 arch/powerpc/kvm/book3s_64_mmu_host.c pte->pagesize, pte->pagesize, pte 194 arch/powerpc/kvm/book3s_64_mmu_host.c cpte->pte = *orig_pte; pte 212 arch/powerpc/kvm/book3s_64_mmu_host.c void kvmppc_mmu_unmap_page(struct kvm_vcpu *vcpu, struct kvmppc_pte *pte) pte 217 arch/powerpc/kvm/book3s_64_mmu_host.c vcpu->arch.mmu.esid_to_vsid(vcpu, pte->eaddr >> SID_SHIFT, &vsid); pte 220 arch/powerpc/kvm/book3s_64_mmu_host.c kvmppc_mmu_pte_vflush(vcpu, pte->vpage, mask); pte 617 arch/powerpc/kvm/book3s_64_mmu_hv.c pte_t *ptep, pte; pte 627 arch/powerpc/kvm/book3s_64_mmu_hv.c pte = kvmppc_read_update_linux_pte(ptep, 1); pte 628 arch/powerpc/kvm/book3s_64_mmu_hv.c if (__pte_write(pte)) pte 1031 arch/powerpc/kvm/book3s_64_mmu_hv.c void kvm_set_spte_hva_hv(struct kvm *kvm, unsigned long hva, pte_t pte) pte 134 arch/powerpc/kvm/book3s_64_mmu_radix.c u64 pte, base, gpa; pte 169 arch/powerpc/kvm/book3s_64_mmu_radix.c pte = __be64_to_cpu(rpte); pte 170 arch/powerpc/kvm/book3s_64_mmu_radix.c if (!(pte & _PAGE_PRESENT)) pte 173 arch/powerpc/kvm/book3s_64_mmu_radix.c if (pte & _PAGE_PTE) pte 176 arch/powerpc/kvm/book3s_64_mmu_radix.c base = pte & RPDB_MASK; pte 177 arch/powerpc/kvm/book3s_64_mmu_radix.c bits = pte & RPDS_MASK; pte 186 arch/powerpc/kvm/book3s_64_mmu_radix.c gpa = pte & 0x01fffffffffff000ul; pte 200 arch/powerpc/kvm/book3s_64_mmu_radix.c gpte->may_read = !!(pte & _PAGE_READ); pte 201 arch/powerpc/kvm/book3s_64_mmu_radix.c gpte->may_write = !!(pte & _PAGE_WRITE); pte 202 arch/powerpc/kvm/book3s_64_mmu_radix.c gpte->may_execute = !!(pte & _PAGE_EXEC); pte 204 arch/powerpc/kvm/book3s_64_mmu_radix.c gpte->rc = pte & (_PAGE_ACCESSED | _PAGE_DIRTY); pte 207 arch/powerpc/kvm/book3s_64_mmu_radix.c *pte_ret_p = pte; pte 253 arch/powerpc/kvm/book3s_64_mmu_radix.c u64 pte; pte 269 arch/powerpc/kvm/book3s_64_mmu_radix.c vcpu->kvm->arch.process_table, pid, &pte); pte 275 arch/powerpc/kvm/book3s_64_mmu_radix.c if (pte & _PAGE_PRIVILEGED) { pte 281 arch/powerpc/kvm/book3s_64_mmu_radix.c if (!(pte & _PAGE_PRIVILEGED)) { pte 346 arch/powerpc/kvm/book3s_64_mmu_radix.c pte_t *ptep, pte_t pte) pte 348 arch/powerpc/kvm/book3s_64_mmu_radix.c radix__set_pte_at(kvm->mm, addr, ptep, pte, 0); pte 375 arch/powerpc/kvm/book3s_64_mmu_radix.c void kvmppc_unmap_pte(struct kvm *kvm, pte_t *pte, unsigned long gpa, pte 386 arch/powerpc/kvm/book3s_64_mmu_radix.c old = kvmppc_radix_update_pte(kvm, pte, ~0UL, 0, gpa, shift); pte 424 arch/powerpc/kvm/book3s_64_mmu_radix.c static void kvmppc_unmap_free_pte(struct kvm *kvm, pte_t *pte, bool full, pte 428 arch/powerpc/kvm/book3s_64_mmu_radix.c memset(pte, 0, sizeof(long) << PTE_INDEX_SIZE); pte 430 arch/powerpc/kvm/book3s_64_mmu_radix.c pte_t *p = pte; pte 443 arch/powerpc/kvm/book3s_64_mmu_radix.c kvmppc_pte_free(pte); pte 465 arch/powerpc/kvm/book3s_64_mmu_radix.c pte_t *pte; pte 467 arch/powerpc/kvm/book3s_64_mmu_radix.c pte = pte_offset_map(p, 0); pte 468 arch/powerpc/kvm/book3s_64_mmu_radix.c kvmppc_unmap_free_pte(kvm, pte, full, lpid); pte 525 arch/powerpc/kvm/book3s_64_mmu_radix.c pte_t *pte = pte_offset_kernel(pmd, 0); pte 535 arch/powerpc/kvm/book3s_64_mmu_radix.c kvmppc_unmap_free_pte(kvm, pte, false, lpid); pte 563 arch/powerpc/kvm/book3s_64_mmu_radix.c int kvmppc_create_pte(struct kvm *kvm, pgd_t *pgtable, pte_t pte, pte 611 arch/powerpc/kvm/book3s_64_mmu_radix.c if (pud_raw(*pud) == pte_raw(pte)) { pte 616 arch/powerpc/kvm/book3s_64_mmu_radix.c WARN_ON_ONCE((pud_val(*pud) ^ pte_val(pte)) & pte 619 arch/powerpc/kvm/book3s_64_mmu_radix.c 0, pte_val(pte), hgpa, PUD_SHIFT); pte 644 arch/powerpc/kvm/book3s_64_mmu_radix.c kvmppc_radix_set_pte_at(kvm, gpa, (pte_t *)pud, pte); pte 662 arch/powerpc/kvm/book3s_64_mmu_radix.c if (pmd_raw(*pmd) == pte_raw(pte)) { pte 667 arch/powerpc/kvm/book3s_64_mmu_radix.c WARN_ON_ONCE((pmd_val(*pmd) ^ pte_val(pte)) & pte 670 arch/powerpc/kvm/book3s_64_mmu_radix.c 0, pte_val(pte), lgpa, PMD_SHIFT); pte 696 arch/powerpc/kvm/book3s_64_mmu_radix.c kvmppc_radix_set_pte_at(kvm, gpa, pmdp_ptep(pmd), pte); pte 711 arch/powerpc/kvm/book3s_64_mmu_radix.c if (pte_raw(*ptep) == pte_raw(pte)) { pte 716 arch/powerpc/kvm/book3s_64_mmu_radix.c WARN_ON_ONCE((pte_val(*ptep) ^ pte_val(pte)) & pte 718 arch/powerpc/kvm/book3s_64_mmu_radix.c kvmppc_radix_update_pte(kvm, ptep, 0, pte_val(pte), gpa, 0); pte 722 arch/powerpc/kvm/book3s_64_mmu_radix.c kvmppc_radix_set_pte_at(kvm, gpa, ptep, pte); pte 778 arch/powerpc/kvm/book3s_64_mmu_radix.c pte_t pte, *ptep; pte 828 arch/powerpc/kvm/book3s_64_mmu_radix.c pte = *ptep; pte 852 arch/powerpc/kvm/book3s_64_mmu_radix.c pte = __pte(pte_val(pte) | (hva & rpnmask)); pte 856 arch/powerpc/kvm/book3s_64_mmu_radix.c pte = __pte(pte_val(pte) | _PAGE_EXEC | _PAGE_ACCESSED); pte 858 arch/powerpc/kvm/book3s_64_mmu_radix.c if (pte_val(pte) & _PAGE_WRITE) pte 859 arch/powerpc/kvm/book3s_64_mmu_radix.c pte = __pte(pte_val(pte) | _PAGE_DIRTY); pte 861 arch/powerpc/kvm/book3s_64_mmu_radix.c pte = __pte(pte_val(pte) & ~(_PAGE_WRITE | _PAGE_DIRTY)); pte 865 arch/powerpc/kvm/book3s_64_mmu_radix.c ret = kvmppc_create_pte(kvm, kvm->arch.pgtable, pte, gpa, level, pte 868 arch/powerpc/kvm/book3s_64_mmu_radix.c *inserted_pte = pte; pte 873 arch/powerpc/kvm/book3s_64_mmu_radix.c if (!ret && (pte_val(pte) & _PAGE_WRITE)) pte 1204 arch/powerpc/kvm/book3s_64_mmu_radix.c unsigned long pte; pte 1285 arch/powerpc/kvm/book3s_64_mmu_radix.c pte = pud_val(pud); pte 1297 arch/powerpc/kvm/book3s_64_mmu_radix.c pte = pmd_val(pmd); pte 1303 arch/powerpc/kvm/book3s_64_mmu_radix.c pte = pte_val(READ_ONCE(*ptep)); pte 1304 arch/powerpc/kvm/book3s_64_mmu_radix.c if (!(pte & _PAGE_PRESENT)) { pte 1311 arch/powerpc/kvm/book3s_64_mmu_radix.c " %lx: %lx %d\n", gpa, pte, shift); pte 444 arch/powerpc/kvm/book3s_64_vio_hv.c pte_t *ptep, pte; pte 459 arch/powerpc/kvm/book3s_64_vio_hv.c pte = *ptep; pte 468 arch/powerpc/kvm/book3s_64_vio_hv.c if (!pte_young(pte)) pte 471 arch/powerpc/kvm/book3s_64_vio_hv.c *phpa = (pte_pfn(pte) << PAGE_SHIFT) | (ua & ((1ULL << shift) - 1)) | pte 1268 arch/powerpc/kvm/book3s_hv_nested.c pte_t pte, *pte_p; pte 1363 arch/powerpc/kvm/book3s_hv_nested.c pte = __pte(0); pte 1369 arch/powerpc/kvm/book3s_hv_nested.c pte = *pte_p; pte 1372 arch/powerpc/kvm/book3s_hv_nested.c if (!pte_present(pte) || (writing && !(pte_val(pte) & _PAGE_WRITE))) { pte 1375 arch/powerpc/kvm/book3s_hv_nested.c writing, kvm_ro, &pte, &level); pte 1394 arch/powerpc/kvm/book3s_hv_nested.c pte = __pte(pte_val(pte) & ~perm); pte 1403 arch/powerpc/kvm/book3s_hv_nested.c pte = __pte(pte_val(pte) | (gpa & mask)); pte 1417 arch/powerpc/kvm/book3s_hv_nested.c ret = kvmppc_create_pte(kvm, gp->shadow_pgtable, pte, n_gpa, level, pte 263 arch/powerpc/kvm/book3s_hv_rm_mmu.c pte_t pte; pte 279 arch/powerpc/kvm/book3s_hv_rm_mmu.c pte = kvmppc_read_update_linux_pte(ptep, writing); pte 280 arch/powerpc/kvm/book3s_hv_rm_mmu.c if (pte_present(pte) && !pte_protnone(pte)) { pte 281 arch/powerpc/kvm/book3s_hv_rm_mmu.c if (writing && !__pte_write(pte)) pte 284 arch/powerpc/kvm/book3s_hv_rm_mmu.c is_ci = pte_ci(pte); pte 285 arch/powerpc/kvm/book3s_hv_rm_mmu.c pa = pte_pfn(pte) << PAGE_SHIFT; pte 338 arch/powerpc/kvm/book3s_hv_rm_mmu.c u64 pte; pte 341 arch/powerpc/kvm/book3s_hv_rm_mmu.c pte = be64_to_cpu(hpte[0]); pte 342 arch/powerpc/kvm/book3s_hv_rm_mmu.c if (!(pte & (HPTE_V_VALID | HPTE_V_ABSENT))) pte 344 arch/powerpc/kvm/book3s_hv_rm_mmu.c __unlock_hpte(hpte, pte); pte 356 arch/powerpc/kvm/book3s_hv_rm_mmu.c u64 pte; pte 360 arch/powerpc/kvm/book3s_hv_rm_mmu.c pte = be64_to_cpu(hpte[0]); pte 361 arch/powerpc/kvm/book3s_hv_rm_mmu.c if (pte & (HPTE_V_VALID | HPTE_V_ABSENT)) { pte 362 arch/powerpc/kvm/book3s_hv_rm_mmu.c __unlock_hpte(hpte, pte); pte 505 arch/powerpc/kvm/book3s_hv_rm_mmu.c u64 pte, orig_pte, pte_r; pte 514 arch/powerpc/kvm/book3s_hv_rm_mmu.c pte = orig_pte = be64_to_cpu(hpte[0]); pte 517 arch/powerpc/kvm/book3s_hv_rm_mmu.c pte = hpte_new_to_old_v(pte, pte_r); pte 520 arch/powerpc/kvm/book3s_hv_rm_mmu.c if ((pte & (HPTE_V_ABSENT | HPTE_V_VALID)) == 0 || pte 521 arch/powerpc/kvm/book3s_hv_rm_mmu.c ((flags & H_AVPN) && (pte & ~0x7fUL) != avpn) || pte 522 arch/powerpc/kvm/book3s_hv_rm_mmu.c ((flags & H_ANDCOND) && (pte & avpn) != 0)) { pte 528 arch/powerpc/kvm/book3s_hv_rm_mmu.c v = pte & ~HPTE_V_HVLOCK; pte 899 arch/powerpc/kvm/book3s_hv_rm_mmu.c pte_t *ptep, pte; pte 914 arch/powerpc/kvm/book3s_hv_rm_mmu.c pte = kvmppc_read_update_linux_pte(ptep, writing); pte 915 arch/powerpc/kvm/book3s_hv_rm_mmu.c if (!pte_present(pte)) pte 921 arch/powerpc/kvm/book3s_hv_rm_mmu.c pa = pte_pfn(pte) << PAGE_SHIFT; pte 56 arch/powerpc/kvm/book3s_mmu_hpte.c void kvmppc_mmu_hpte_cache_map(struct kvm_vcpu *vcpu, struct hpte_cache *pte) pte 61 arch/powerpc/kvm/book3s_mmu_hpte.c trace_kvm_book3s_mmu_map(pte); pte 66 arch/powerpc/kvm/book3s_mmu_hpte.c index = kvmppc_mmu_hash_pte(pte->pte.eaddr); pte 67 arch/powerpc/kvm/book3s_mmu_hpte.c hlist_add_head_rcu(&pte->list_pte, &vcpu3s->hpte_hash_pte[index]); pte 70 arch/powerpc/kvm/book3s_mmu_hpte.c index = kvmppc_mmu_hash_pte_long(pte->pte.eaddr); pte 71 arch/powerpc/kvm/book3s_mmu_hpte.c hlist_add_head_rcu(&pte->list_pte_long, pte 75 arch/powerpc/kvm/book3s_mmu_hpte.c index = kvmppc_mmu_hash_vpte(pte->pte.vpage); pte 76 arch/powerpc/kvm/book3s_mmu_hpte.c hlist_add_head_rcu(&pte->list_vpte, &vcpu3s->hpte_hash_vpte[index]); pte 79 arch/powerpc/kvm/book3s_mmu_hpte.c index = kvmppc_mmu_hash_vpte_long(pte->pte.vpage); pte 80 arch/powerpc/kvm/book3s_mmu_hpte.c hlist_add_head_rcu(&pte->list_vpte_long, pte 85 arch/powerpc/kvm/book3s_mmu_hpte.c index = kvmppc_mmu_hash_vpte_64k(pte->pte.vpage); pte 86 arch/powerpc/kvm/book3s_mmu_hpte.c hlist_add_head_rcu(&pte->list_vpte_64k, pte 97 arch/powerpc/kvm/book3s_mmu_hpte.c struct hpte_cache *pte = container_of(head, struct hpte_cache, rcu_head); pte 98 arch/powerpc/kvm/book3s_mmu_hpte.c kmem_cache_free(hpte_cache, pte); pte 101 arch/powerpc/kvm/book3s_mmu_hpte.c static void invalidate_pte(struct kvm_vcpu *vcpu, struct hpte_cache *pte) pte 105 arch/powerpc/kvm/book3s_mmu_hpte.c trace_kvm_book3s_mmu_invalidate(pte); pte 108 arch/powerpc/kvm/book3s_mmu_hpte.c kvmppc_mmu_invalidate_pte(vcpu, pte); pte 113 arch/powerpc/kvm/book3s_mmu_hpte.c if (hlist_unhashed(&pte->list_pte)) { pte 118 arch/powerpc/kvm/book3s_mmu_hpte.c hlist_del_init_rcu(&pte->list_pte); pte 119 arch/powerpc/kvm/book3s_mmu_hpte.c hlist_del_init_rcu(&pte->list_pte_long); pte 120 arch/powerpc/kvm/book3s_mmu_hpte.c hlist_del_init_rcu(&pte->list_vpte); pte 121 arch/powerpc/kvm/book3s_mmu_hpte.c hlist_del_init_rcu(&pte->list_vpte_long); pte 123 arch/powerpc/kvm/book3s_mmu_hpte.c hlist_del_init_rcu(&pte->list_vpte_64k); pte 129 arch/powerpc/kvm/book3s_mmu_hpte.c call_rcu(&pte->rcu_head, free_pte_rcu); pte 135 arch/powerpc/kvm/book3s_mmu_hpte.c struct hpte_cache *pte; pte 143 arch/powerpc/kvm/book3s_mmu_hpte.c hlist_for_each_entry_rcu(pte, list, list_vpte_long) pte 144 arch/powerpc/kvm/book3s_mmu_hpte.c invalidate_pte(vcpu, pte); pte 154 arch/powerpc/kvm/book3s_mmu_hpte.c struct hpte_cache *pte; pte 162 arch/powerpc/kvm/book3s_mmu_hpte.c hlist_for_each_entry_rcu(pte, list, list_pte) pte 163 arch/powerpc/kvm/book3s_mmu_hpte.c if ((pte->pte.eaddr & ~0xfffUL) == guest_ea) pte 164 arch/powerpc/kvm/book3s_mmu_hpte.c invalidate_pte(vcpu, pte); pte 173 arch/powerpc/kvm/book3s_mmu_hpte.c struct hpte_cache *pte; pte 182 arch/powerpc/kvm/book3s_mmu_hpte.c hlist_for_each_entry_rcu(pte, list, list_pte_long) pte 183 arch/powerpc/kvm/book3s_mmu_hpte.c if ((pte->pte.eaddr & 0x0ffff000UL) == guest_ea) pte 184 arch/powerpc/kvm/book3s_mmu_hpte.c invalidate_pte(vcpu, pte); pte 216 arch/powerpc/kvm/book3s_mmu_hpte.c struct hpte_cache *pte; pte 224 arch/powerpc/kvm/book3s_mmu_hpte.c hlist_for_each_entry_rcu(pte, list, list_vpte) pte 225 arch/powerpc/kvm/book3s_mmu_hpte.c if ((pte->pte.vpage & vp_mask) == guest_vp) pte 226 arch/powerpc/kvm/book3s_mmu_hpte.c invalidate_pte(vcpu, pte); pte 237 arch/powerpc/kvm/book3s_mmu_hpte.c struct hpte_cache *pte; pte 246 arch/powerpc/kvm/book3s_mmu_hpte.c hlist_for_each_entry_rcu(pte, list, list_vpte_64k) pte 247 arch/powerpc/kvm/book3s_mmu_hpte.c if ((pte->pte.vpage & vp_mask) == guest_vp) pte 248 arch/powerpc/kvm/book3s_mmu_hpte.c invalidate_pte(vcpu, pte); pte 259 arch/powerpc/kvm/book3s_mmu_hpte.c struct hpte_cache *pte; pte 268 arch/powerpc/kvm/book3s_mmu_hpte.c hlist_for_each_entry_rcu(pte, list, list_vpte_long) pte 269 arch/powerpc/kvm/book3s_mmu_hpte.c if ((pte->pte.vpage & vp_mask) == guest_vp) pte 270 arch/powerpc/kvm/book3s_mmu_hpte.c invalidate_pte(vcpu, pte); pte 301 arch/powerpc/kvm/book3s_mmu_hpte.c struct hpte_cache *pte; pte 311 arch/powerpc/kvm/book3s_mmu_hpte.c hlist_for_each_entry_rcu(pte, list, list_vpte_long) pte 312 arch/powerpc/kvm/book3s_mmu_hpte.c if ((pte->pte.raddr >= pa_start) && pte 313 arch/powerpc/kvm/book3s_mmu_hpte.c (pte->pte.raddr < pa_end)) pte 314 arch/powerpc/kvm/book3s_mmu_hpte.c invalidate_pte(vcpu, pte); pte 323 arch/powerpc/kvm/book3s_mmu_hpte.c struct hpte_cache *pte; pte 328 arch/powerpc/kvm/book3s_mmu_hpte.c pte = kmem_cache_zalloc(hpte_cache, GFP_KERNEL); pte 330 arch/powerpc/kvm/book3s_mmu_hpte.c return pte; pte 333 arch/powerpc/kvm/book3s_mmu_hpte.c void kvmppc_mmu_hpte_cache_free(struct hpte_cache *pte) pte 335 arch/powerpc/kvm/book3s_mmu_hpte.c kmem_cache_free(hpte_cache, pte); pte 443 arch/powerpc/kvm/book3s_pr.c static void kvm_set_spte_hva_pr(struct kvm *kvm, unsigned long hva, pte_t pte) pte 625 arch/powerpc/kvm/book3s_pr.c static void kvmppc_patch_dcbz(struct kvm_vcpu *vcpu, struct kvmppc_pte *pte) pte 632 arch/powerpc/kvm/book3s_pr.c hpage = gfn_to_page(vcpu->kvm, pte->raddr >> PAGE_SHIFT); pte 636 arch/powerpc/kvm/book3s_pr.c hpage_offset = pte->raddr & ~PAGE_MASK; pte 675 arch/powerpc/kvm/book3s_pr.c struct kvmppc_pte pte = { 0 }; pte 686 arch/powerpc/kvm/book3s_pr.c page_found = vcpu->arch.mmu.xlate(vcpu, eaddr, &pte, data, iswrite); pte 688 arch/powerpc/kvm/book3s_pr.c pte.may_execute = true; pte 689 arch/powerpc/kvm/book3s_pr.c pte.may_read = true; pte 690 arch/powerpc/kvm/book3s_pr.c pte.may_write = true; pte 691 arch/powerpc/kvm/book3s_pr.c pte.raddr = eaddr & KVM_PAM; pte 692 arch/powerpc/kvm/book3s_pr.c pte.eaddr = eaddr; pte 693 arch/powerpc/kvm/book3s_pr.c pte.vpage = eaddr >> 12; pte 694 arch/powerpc/kvm/book3s_pr.c pte.page_size = MMU_PAGE_64K; pte 695 arch/powerpc/kvm/book3s_pr.c pte.wimg = HPTE_R_M; pte 700 arch/powerpc/kvm/book3s_pr.c pte.vpage |= ((u64)VSID_REAL << (SID_SHIFT - 12)); pte 705 arch/powerpc/kvm/book3s_pr.c ((pte.raddr & SPLIT_HACK_MASK) == SPLIT_HACK_OFFS)) pte 706 arch/powerpc/kvm/book3s_pr.c pte.raddr &= ~SPLIT_HACK_MASK; pte 712 arch/powerpc/kvm/book3s_pr.c pte.vpage |= ((u64)VSID_REAL_DR << (SID_SHIFT - 12)); pte 714 arch/powerpc/kvm/book3s_pr.c pte.vpage |= ((u64)VSID_REAL_IR << (SID_SHIFT - 12)); pte 715 arch/powerpc/kvm/book3s_pr.c pte.vpage |= vsid; pte 729 arch/powerpc/kvm/book3s_pr.c pte.may_execute = !data; pte 750 arch/powerpc/kvm/book3s_pr.c } else if (kvmppc_visible_gpa(vcpu, pte.raddr)) { pte 757 arch/powerpc/kvm/book3s_pr.c kvmppc_mmu_unmap_page(vcpu, &pte); pte 760 arch/powerpc/kvm/book3s_pr.c if (kvmppc_mmu_map_page(vcpu, &pte, iswrite) == -EIO) { pte 769 arch/powerpc/kvm/book3s_pr.c kvmppc_patch_dcbz(vcpu, &pte); pte 773 arch/powerpc/kvm/book3s_pr.c vcpu->arch.paddr_accessed = pte.raddr; pte 774 arch/powerpc/kvm/book3s_pr.c vcpu->arch.vaddr_accessed = pte.eaddr; pte 92 arch/powerpc/kvm/book3s_pr_papr.c unsigned long pte[2]; pte 98 arch/powerpc/kvm/book3s_pr_papr.c if (copy_from_user(pte, (void __user *)pteg, sizeof(pte))) pte 100 arch/powerpc/kvm/book3s_pr_papr.c pte[0] = be64_to_cpu((__force __be64)pte[0]); pte 101 arch/powerpc/kvm/book3s_pr_papr.c pte[1] = be64_to_cpu((__force __be64)pte[1]); pte 104 arch/powerpc/kvm/book3s_pr_papr.c if ((pte[0] & HPTE_V_VALID) == 0 || pte 105 arch/powerpc/kvm/book3s_pr_papr.c ((flags & H_AVPN) && (pte[0] & ~0x7fUL) != avpn) || pte 106 arch/powerpc/kvm/book3s_pr_papr.c ((flags & H_ANDCOND) && (pte[0] & avpn) != 0)) pte 113 arch/powerpc/kvm/book3s_pr_papr.c rb = compute_tlbie_rb(pte[0], pte[1], pte_index); pte 117 arch/powerpc/kvm/book3s_pr_papr.c kvmppc_set_gpr(vcpu, 4, pte[0]); pte 118 arch/powerpc/kvm/book3s_pr_papr.c kvmppc_set_gpr(vcpu, 5, pte[1]); pte 156 arch/powerpc/kvm/book3s_pr_papr.c unsigned long pte[2]; pte 179 arch/powerpc/kvm/book3s_pr_papr.c if (copy_from_user(pte, (void __user *)pteg, sizeof(pte))) { pte 183 arch/powerpc/kvm/book3s_pr_papr.c pte[0] = be64_to_cpu((__force __be64)pte[0]); pte 184 arch/powerpc/kvm/book3s_pr_papr.c pte[1] = be64_to_cpu((__force __be64)pte[1]); pte 189 arch/powerpc/kvm/book3s_pr_papr.c if ((pte[0] & HPTE_V_VALID) == 0 || pte 190 arch/powerpc/kvm/book3s_pr_papr.c ((flags & H_AVPN) && (pte[0] & ~0x7fUL) != tsl) || pte 191 arch/powerpc/kvm/book3s_pr_papr.c ((flags & H_ANDCOND) && (pte[0] & tsl) != 0)) { pte 200 arch/powerpc/kvm/book3s_pr_papr.c rb = compute_tlbie_rb(pte[0], pte[1], pte 204 arch/powerpc/kvm/book3s_pr_papr.c tsh |= (pte[1] & (HPTE_R_C | HPTE_R_R)) << 43; pte 220 arch/powerpc/kvm/book3s_pr_papr.c unsigned long pte[2]; pte 226 arch/powerpc/kvm/book3s_pr_papr.c if (copy_from_user(pte, (void __user *)pteg, sizeof(pte))) pte 228 arch/powerpc/kvm/book3s_pr_papr.c pte[0] = be64_to_cpu((__force __be64)pte[0]); pte 229 arch/powerpc/kvm/book3s_pr_papr.c pte[1] = be64_to_cpu((__force __be64)pte[1]); pte 232 arch/powerpc/kvm/book3s_pr_papr.c if ((pte[0] & HPTE_V_VALID) == 0 || pte 233 arch/powerpc/kvm/book3s_pr_papr.c ((flags & H_AVPN) && (pte[0] & ~0x7fUL) != avpn)) pte 236 arch/powerpc/kvm/book3s_pr_papr.c v = pte[0]; pte 237 arch/powerpc/kvm/book3s_pr_papr.c r = pte[1]; pte 244 arch/powerpc/kvm/book3s_pr_papr.c pte[1] = r; pte 248 arch/powerpc/kvm/book3s_pr_papr.c pte[0] = (__force u64)cpu_to_be64(pte[0]); pte 249 arch/powerpc/kvm/book3s_pr_papr.c pte[1] = (__force u64)cpu_to_be64(pte[1]); pte 251 arch/powerpc/kvm/book3s_pr_papr.c if (copy_to_user((void __user *)pteg, pte, sizeof(pte))) pte 1964 arch/powerpc/kvm/booke.c enum xlate_readwrite xlrw, struct kvmppc_pte *pte) pte 1972 arch/powerpc/kvm/booke.c pte->eaddr = eaddr; pte 1973 arch/powerpc/kvm/booke.c pte->raddr = (vcpu->arch.magic_page_pa & PAGE_MASK) | pte 1975 arch/powerpc/kvm/booke.c pte->vpage = eaddr >> PAGE_SHIFT; pte 1976 arch/powerpc/kvm/booke.c pte->may_read = true; pte 1977 arch/powerpc/kvm/booke.c pte->may_write = true; pte 1978 arch/powerpc/kvm/booke.c pte->may_execute = true; pte 2002 arch/powerpc/kvm/booke.c pte->eaddr = eaddr; pte 2003 arch/powerpc/kvm/booke.c pte->raddr = (gpaddr & PAGE_MASK) | (eaddr & ~PAGE_MASK); pte 2004 arch/powerpc/kvm/booke.c pte->vpage = eaddr >> PAGE_SHIFT; pte 2007 arch/powerpc/kvm/booke.c pte->may_read = true; pte 2008 arch/powerpc/kvm/booke.c pte->may_write = true; pte 2009 arch/powerpc/kvm/booke.c pte->may_execute = true; pte 479 arch/powerpc/kvm/e500_mmu_host.c pte_t pte = READ_ONCE(*ptep); pte 481 arch/powerpc/kvm/e500_mmu_host.c if (pte_present(pte)) { pte 482 arch/powerpc/kvm/e500_mmu_host.c wimg = (pte_val(pte) >> PTE_WIMGE_SHIFT) & pte 757 arch/powerpc/kvm/e500_mmu_host.c int kvm_set_spte_hva(struct kvm *kvm, unsigned long hva, pte_t pte) pte 327 arch/powerpc/kvm/powerpc.c struct kvmppc_pte pte; pte 340 arch/powerpc/kvm/powerpc.c XLATE_WRITE, &pte); pte 344 arch/powerpc/kvm/powerpc.c *eaddr = pte.raddr; pte 346 arch/powerpc/kvm/powerpc.c if (!pte.may_write) pte 351 arch/powerpc/kvm/powerpc.c ((pte.raddr & KVM_PAM & PAGE_MASK) == mp_pa) && pte 354 arch/powerpc/kvm/powerpc.c magic += pte.eaddr & 0xfff; pte 359 arch/powerpc/kvm/powerpc.c if (kvm_write_guest(vcpu->kvm, pte.raddr, ptr, size)) pte 370 arch/powerpc/kvm/powerpc.c struct kvmppc_pte pte; pte 383 arch/powerpc/kvm/powerpc.c XLATE_READ, &pte); pte 387 arch/powerpc/kvm/powerpc.c *eaddr = pte.raddr; pte 389 arch/powerpc/kvm/powerpc.c if (!pte.may_read) pte 392 arch/powerpc/kvm/powerpc.c if (!data && !pte.may_execute) pte 397 arch/powerpc/kvm/powerpc.c ((pte.raddr & KVM_PAM & PAGE_MASK) == mp_pa) && pte 400 arch/powerpc/kvm/powerpc.c magic += pte.eaddr & 0xfff; pte 405 arch/powerpc/kvm/powerpc.c if (kvm_read_guest(vcpu->kvm, pte.raddr, ptr, size)) pte 64 arch/powerpc/kvm/trace_pr.h TP_PROTO(struct hpte_cache *pte), pte 65 arch/powerpc/kvm/trace_pr.h TP_ARGS(pte), pte 77 arch/powerpc/kvm/trace_pr.h __entry->host_vpn = pte->host_vpn; pte 78 arch/powerpc/kvm/trace_pr.h __entry->pfn = pte->pfn; pte 79 arch/powerpc/kvm/trace_pr.h __entry->eaddr = pte->pte.eaddr; pte 80 arch/powerpc/kvm/trace_pr.h __entry->vpage = pte->pte.vpage; pte 81 arch/powerpc/kvm/trace_pr.h __entry->raddr = pte->pte.raddr; pte 82 arch/powerpc/kvm/trace_pr.h __entry->flags = (pte->pte.may_read ? 0x4 : 0) | pte 83 arch/powerpc/kvm/trace_pr.h (pte->pte.may_write ? 0x2 : 0) | pte 84 arch/powerpc/kvm/trace_pr.h (pte->pte.may_execute ? 0x1 : 0); pte 93 arch/powerpc/kvm/trace_pr.h TP_PROTO(struct hpte_cache *pte), pte 94 arch/powerpc/kvm/trace_pr.h TP_ARGS(pte), pte 106 arch/powerpc/kvm/trace_pr.h __entry->host_vpn = pte->host_vpn; pte 107 arch/powerpc/kvm/trace_pr.h __entry->pfn = pte->pfn; pte 108 arch/powerpc/kvm/trace_pr.h __entry->eaddr = pte->pte.eaddr; pte 109 arch/powerpc/kvm/trace_pr.h __entry->vpage = pte->pte.vpage; pte 110 arch/powerpc/kvm/trace_pr.h __entry->raddr = pte->pte.raddr; pte 111 arch/powerpc/kvm/trace_pr.h __entry->flags = (pte->pte.may_read ? 0x4 : 0) | pte 112 arch/powerpc/kvm/trace_pr.h (pte->pte.may_write ? 0x2 : 0) | pte 113 arch/powerpc/kvm/trace_pr.h (pte->pte.may_execute ? 0x1 : 0); pte 34 arch/powerpc/mm/book3s64/hash_4k.c pte_t pte = READ_ONCE(*ptep); pte 36 arch/powerpc/mm/book3s64/hash_4k.c old_pte = pte_val(pte); pte 52 arch/powerpc/mm/book3s64/hash_64k.c pte_t pte = READ_ONCE(*ptep); pte 54 arch/powerpc/mm/book3s64/hash_64k.c old_pte = pte_val(pte); pte 238 arch/powerpc/mm/book3s64/hash_64k.c pte_t pte = READ_ONCE(*ptep); pte 240 arch/powerpc/mm/book3s64/hash_64k.c old_pte = pte_val(pte); pte 252 arch/powerpc/mm/book3s64/hash_64k.c unlikely(pte_ci(pte))) pte 148 arch/powerpc/mm/book3s64/hash_hugetlbpage.c pte_t *ptep, pte_t old_pte, pte_t pte) pte 153 arch/powerpc/mm/book3s64/hash_hugetlbpage.c old_pte, pte); pte 154 arch/powerpc/mm/book3s64/hash_hugetlbpage.c set_huge_pte_at(vma->vm_mm, addr, ptep, pte); pte 810 arch/powerpc/mm/book3s64/hash_native.c real_pte_t pte; pte 824 arch/powerpc/mm/book3s64/hash_native.c pte = batch->pte[i]; pte 826 arch/powerpc/mm/book3s64/hash_native.c pte_iterate_hashed_subpages(pte, psize, vpn, index, shift) { pte 828 arch/powerpc/mm/book3s64/hash_native.c hidx = __rpte_to_hidx(pte, index); pte 855 arch/powerpc/mm/book3s64/hash_native.c pte = batch->pte[i]; pte 857 arch/powerpc/mm/book3s64/hash_native.c pte_iterate_hashed_subpages(pte, psize, pte 872 arch/powerpc/mm/book3s64/hash_native.c pte = batch->pte[i]; pte 874 arch/powerpc/mm/book3s64/hash_native.c pte_iterate_hashed_subpages(pte, psize, pte 42 arch/powerpc/mm/book3s64/hash_tlb.c pte_t *ptep, unsigned long pte, int huge) pte 73 arch/powerpc/mm/book3s64/hash_tlb.c psize = pte_pagesize_index(mm, addr, pte); /* shutup gcc */ pte 76 arch/powerpc/mm/book3s64/hash_tlb.c psize = pte_pagesize_index(mm, addr, pte); pte 98 arch/powerpc/mm/book3s64/hash_tlb.c rpte = __real_pte(__pte(pte), ptep, offset); pte 130 arch/powerpc/mm/book3s64/hash_tlb.c batch->pte[i] = rpte; pte 152 arch/powerpc/mm/book3s64/hash_tlb.c flush_hash_page(batch->vpn[0], batch->pte[0], pte 217 arch/powerpc/mm/book3s64/hash_tlb.c unsigned long pte; pte 221 arch/powerpc/mm/book3s64/hash_tlb.c pte = pte_val(*ptep); pte 223 arch/powerpc/mm/book3s64/hash_tlb.c trace_hugepage_invalidate(start, pte); pte 224 arch/powerpc/mm/book3s64/hash_tlb.c if (!(pte & H_PAGE_HASHPTE)) pte 227 arch/powerpc/mm/book3s64/hash_tlb.c hpte_do_hugepage_flush(mm, start, (pmd_t *)ptep, pte); pte 229 arch/powerpc/mm/book3s64/hash_tlb.c hpte_need_flush(mm, start, ptep, pte, hugepage_shift); pte 237 arch/powerpc/mm/book3s64/hash_tlb.c pte_t *pte; pte 253 arch/powerpc/mm/book3s64/hash_tlb.c for (pte = start_pte; pte < start_pte + PTRS_PER_PTE; pte++) { pte 254 arch/powerpc/mm/book3s64/hash_tlb.c unsigned long pteval = pte_val(*pte); pte 256 arch/powerpc/mm/book3s64/hash_tlb.c hpte_need_flush(mm, addr, pte, pteval, 0); pte 1093 arch/powerpc/mm/book3s64/hash_utils.c unsigned int hash_page_do_lazy_icache(unsigned int pp, pte_t pte, int trap) pte 1097 arch/powerpc/mm/book3s64/hash_utils.c if (!pfn_valid(pte_pfn(pte))) pte 1100 arch/powerpc/mm/book3s64/hash_utils.c page = pte_page(pte); pte 1210 arch/powerpc/mm/book3s64/hash_utils.c int ssize, int psize, int lpsize, unsigned long pte) pte 1217 arch/powerpc/mm/book3s64/hash_utils.c trap, vsid, ssize, psize, lpsize, pte); pte 1722 arch/powerpc/mm/book3s64/hash_utils.c void flush_hash_page(unsigned long vpn, real_pte_t pte, int psize, int ssize, pte 1729 arch/powerpc/mm/book3s64/hash_utils.c pte_iterate_hashed_subpages(pte, psize, vpn, index, shift) { pte 1730 arch/powerpc/mm/book3s64/hash_utils.c gslot = pte_get_hash_gslot(vpn, shift, ssize, pte, index); pte 1811 arch/powerpc/mm/book3s64/hash_utils.c flush_hash_page(batch->vpn[i], batch->pte[i], pte 437 arch/powerpc/mm/book3s64/pgtable.c pte_t *ptep, pte_t old_pte, pte_t pte) pte 441 arch/powerpc/mm/book3s64/pgtable.c ptep, old_pte, pte); pte 442 arch/powerpc/mm/book3s64/pgtable.c set_pte_at(vma->vm_mm, addr, ptep, pte); pte 378 arch/powerpc/mm/book3s64/pkeys.c bool arch_pte_access_permitted(u64 pte, bool write, bool execute) pte 383 arch/powerpc/mm/book3s64/pkeys.c return pkey_access_permitted(pte_to_pkey_bits(pte), write, execute); pte 97 arch/powerpc/mm/book3s64/radix_hugetlbpage.c pte_t old_pte, pte_t pte) pte 105 arch/powerpc/mm/book3s64/radix_hugetlbpage.c if (is_pte_rw_upgrade(pte_val(old_pte), pte_val(pte)) && pte 109 arch/powerpc/mm/book3s64/radix_hugetlbpage.c set_huge_pte_at(vma->vm_mm, addr, ptep, pte); pte 664 arch/powerpc/mm/book3s64/radix_pgtable.c pte_t *pte; pte 668 arch/powerpc/mm/book3s64/radix_pgtable.c pte = pte_start + i; pte 669 arch/powerpc/mm/book3s64/radix_pgtable.c if (!pte_none(*pte)) pte 693 arch/powerpc/mm/book3s64/radix_pgtable.c pte_t *pte; pte 709 arch/powerpc/mm/book3s64/radix_pgtable.c pte_clear(&init_mm, params->aligned_start, params->pte); pte 720 arch/powerpc/mm/book3s64/radix_pgtable.c pte_t *pte; pte 722 arch/powerpc/mm/book3s64/radix_pgtable.c pte = pte_start + pte_index(addr); pte 723 arch/powerpc/mm/book3s64/radix_pgtable.c for (; addr < end; addr = next, pte++) { pte 728 arch/powerpc/mm/book3s64/radix_pgtable.c if (!pte_present(*pte)) pte 740 arch/powerpc/mm/book3s64/radix_pgtable.c pte_clear(&init_mm, addr, pte); pte 748 arch/powerpc/mm/book3s64/radix_pgtable.c unsigned long size, pte_t *pte) pte 777 arch/powerpc/mm/book3s64/radix_pgtable.c params.pte = pte; pte 787 arch/powerpc/mm/book3s64/radix_pgtable.c pte_clear(&init_mm, addr, pte); pte 1068 arch/powerpc/mm/book3s64/radix_pgtable.c pte_t old_pte, pte_t pte) pte 1077 arch/powerpc/mm/book3s64/radix_pgtable.c if (is_pte_rw_upgrade(pte_val(old_pte), pte_val(pte)) && pte 1081 arch/powerpc/mm/book3s64/radix_pgtable.c set_pte_at(mm, addr, ptep, pte); pte 1135 arch/powerpc/mm/book3s64/radix_pgtable.c pte_t *pte; pte 1136 arch/powerpc/mm/book3s64/radix_pgtable.c pte = (pte_t *)pmd_page_vaddr(pmd[i]); pte 1138 arch/powerpc/mm/book3s64/radix_pgtable.c pte_free_kernel(&init_mm, pte); pte 1172 arch/powerpc/mm/book3s64/radix_pgtable.c pte_t *pte; pte 1174 arch/powerpc/mm/book3s64/radix_pgtable.c pte = (pte_t *)pmd_page_vaddr(*pmd); pte 1179 arch/powerpc/mm/book3s64/radix_pgtable.c pte_free_kernel(&init_mm, pte); pte 59 arch/powerpc/mm/book3s64/subpage_prot.c pte_t *pte; pte 71 arch/powerpc/mm/book3s64/subpage_prot.c pte = pte_offset_map_lock(mm, pmd, addr, &ptl); pte 74 arch/powerpc/mm/book3s64/subpage_prot.c pte_update(mm, addr, pte, 0, 0, 0); pte 76 arch/powerpc/mm/book3s64/subpage_prot.c ++pte; pte 79 arch/powerpc/mm/book3s64/subpage_prot.c pte_unmap_unlock(pte - 1, ptl); pte 45 arch/powerpc/mm/ioremap.c pte_t pte = __pte(flags); pte 49 arch/powerpc/mm/ioremap.c if (pte_write(pte)) pte 50 arch/powerpc/mm/ioremap.c pte = pte_mkdirty(pte); pte 53 arch/powerpc/mm/ioremap.c pte = pte_exprotect(pte); pte 54 arch/powerpc/mm/ioremap.c pte = pte_mkprivileged(pte); pte 57 arch/powerpc/mm/ioremap.c return iowa_ioremap(addr, size, pte_pgprot(pte), caller); pte 58 arch/powerpc/mm/ioremap.c return __ioremap_caller(addr, size, pte_pgprot(pte), caller); pte 99 arch/powerpc/mm/kasan/kasan_init_32.c pte_t pte = pfn_pte(PHYS_PFN(__pa(va)), PAGE_KERNEL); pte 104 arch/powerpc/mm/kasan/kasan_init_32.c __set_pte_at(&init_mm, k_cur, pte_offset_kernel(pmd, k_cur), pte, 0); pte 126 arch/powerpc/mm/nohash/book3e_hugetlbpage.c book3e_hugetlb_preload(struct vm_area_struct *vma, unsigned long ea, pte_t pte) pte 163 arch/powerpc/mm/nohash/book3e_hugetlbpage.c mas2 |= (pte_val(pte) >> PTE_WIMGE_SHIFT) & MAS2_WIMGE_MASK; pte 164 arch/powerpc/mm/nohash/book3e_hugetlbpage.c mas7_3 = (u64)pte_pfn(pte) << PAGE_SHIFT; pte 165 arch/powerpc/mm/nohash/book3e_hugetlbpage.c mas7_3 |= (pte_val(pte) >> PTE_BAP_SHIFT) & MAS3_BAP_MASK; pte 166 arch/powerpc/mm/nohash/book3e_hugetlbpage.c if (!pte_dirty(pte)) pte 100 arch/powerpc/mm/pgtable-frag.c pte_t *pte; pte 102 arch/powerpc/mm/pgtable-frag.c pte = get_pte_from_cache(mm); pte 103 arch/powerpc/mm/pgtable-frag.c if (pte) pte 104 arch/powerpc/mm/pgtable-frag.c return pte; pte 41 arch/powerpc/mm/pgtable.c static inline int pte_looks_normal(pte_t pte) pte 44 arch/powerpc/mm/pgtable.c if (pte_present(pte) && !pte_special(pte)) { pte 45 arch/powerpc/mm/pgtable.c if (pte_ci(pte)) pte 47 arch/powerpc/mm/pgtable.c if (pte_user(pte)) pte 53 arch/powerpc/mm/pgtable.c static struct page *maybe_pte_to_page(pte_t pte) pte 55 arch/powerpc/mm/pgtable.c unsigned long pfn = pte_pfn(pte); pte 74 arch/powerpc/mm/pgtable.c static pte_t set_pte_filter_hash(pte_t pte) pte 77 arch/powerpc/mm/pgtable.c return pte; pte 79 arch/powerpc/mm/pgtable.c pte = __pte(pte_val(pte) & ~_PAGE_HPTEFLAGS); pte 80 arch/powerpc/mm/pgtable.c if (pte_looks_normal(pte) && !(cpu_has_feature(CPU_FTR_COHERENT_ICACHE) || pte 82 arch/powerpc/mm/pgtable.c struct page *pg = maybe_pte_to_page(pte); pte 84 arch/powerpc/mm/pgtable.c return pte; pte 90 arch/powerpc/mm/pgtable.c return pte; pte 95 arch/powerpc/mm/pgtable.c static pte_t set_pte_filter_hash(pte_t pte) { return pte; } pte 103 arch/powerpc/mm/pgtable.c static pte_t set_pte_filter(pte_t pte) pte 108 arch/powerpc/mm/pgtable.c return set_pte_filter_hash(pte); pte 111 arch/powerpc/mm/pgtable.c if (!pte_exec(pte) || !pte_looks_normal(pte)) pte 112 arch/powerpc/mm/pgtable.c return pte; pte 115 arch/powerpc/mm/pgtable.c pg = maybe_pte_to_page(pte); pte 117 arch/powerpc/mm/pgtable.c return pte; pte 121 arch/powerpc/mm/pgtable.c return pte; pte 127 arch/powerpc/mm/pgtable.c return pte; pte 131 arch/powerpc/mm/pgtable.c return pte_exprotect(pte); pte 134 arch/powerpc/mm/pgtable.c static pte_t set_access_flags_filter(pte_t pte, struct vm_area_struct *vma, pte 140 arch/powerpc/mm/pgtable.c return pte; pte 147 arch/powerpc/mm/pgtable.c if (dirty || pte_exec(pte) || !is_exec_fault()) pte 148 arch/powerpc/mm/pgtable.c return pte; pte 156 arch/powerpc/mm/pgtable.c return pte; pte 160 arch/powerpc/mm/pgtable.c pg = maybe_pte_to_page(pte); pte 173 arch/powerpc/mm/pgtable.c return pte_mkexec(pte); pte 180 arch/powerpc/mm/pgtable.c pte_t pte) pte 189 arch/powerpc/mm/pgtable.c pte = pte_mkpte(pte); pte 195 arch/powerpc/mm/pgtable.c pte = set_pte_filter(pte); pte 198 arch/powerpc/mm/pgtable.c __set_pte_at(mm, addr, ptep, pte, 0); pte 225 arch/powerpc/mm/pgtable.c pte_t pte, int dirty) pte 233 arch/powerpc/mm/pgtable.c ptep_set_access_flags(vma, addr, ptep, pte, dirty); pte 238 arch/powerpc/mm/pgtable.c pte = set_access_flags_filter(pte, vma, dirty); pte 239 arch/powerpc/mm/pgtable.c changed = !pte_same(*(ptep), pte); pte 257 arch/powerpc/mm/pgtable.c __ptep_set_access_flags(vma, ptep, pte, addr, psize); pte 138 arch/powerpc/mm/pgtable_32.c pte_t *pte; pte 147 arch/powerpc/mm/pgtable_32.c pte = pte_offset_map(pmd, addr & PAGE_MASK); pte 148 arch/powerpc/mm/pgtable_32.c if (pte) { pte 150 arch/powerpc/mm/pgtable_32.c *ptep = pte; pte 156 arch/powerpc/mm/ptdump/hashpagetable.c *flag, u64 pte, int num) pte 169 arch/powerpc/mm/ptdump/hashpagetable.c val = pte & flag->val; pte 174 arch/powerpc/mm/ptdump/hashpagetable.c if ((pte & flag->mask) == flag->val) pte 374 arch/powerpc/mm/ptdump/hashpagetable.c pte_t *pte = pte_offset_kernel(pmd, 0); pte 378 arch/powerpc/mm/ptdump/hashpagetable.c for (i = 0; i < PTRS_PER_PTE; i++, pte++) { pte 380 arch/powerpc/mm/ptdump/hashpagetable.c pteval = pte_val(*pte); pte 115 arch/powerpc/mm/ptdump/ptdump.c *flag, u64 pte, int num) pte 128 arch/powerpc/mm/ptdump/ptdump.c val = pte & flag->val; pte 133 arch/powerpc/mm/ptdump/ptdump.c if ((pte & flag->mask) == flag->val) pte 177 arch/powerpc/mm/ptdump/ptdump.c pte_t pte = __pte(st->current_flags); pte 182 arch/powerpc/mm/ptdump/ptdump.c if (!pte_write(pte) || !pte_exec(pte)) pte 254 arch/powerpc/mm/ptdump/ptdump.c pte_t *pte = pte_offset_kernel(pmd, 0); pte 258 arch/powerpc/mm/ptdump/ptdump.c for (i = 0; i < PTRS_PER_PTE; i++, pte++) { pte 260 arch/powerpc/mm/ptdump/ptdump.c note_page(st, addr, 4, pte_val(*pte), PAGE_SIZE); pte 115 arch/powerpc/perf/callchain.c pte_t *ptep, pte; pte 136 arch/powerpc/perf/callchain.c pte = READ_ONCE(*ptep); pte 137 arch/powerpc/perf/callchain.c if (!pte_present(pte) || !pte_user(pte)) pte 139 arch/powerpc/perf/callchain.c pfn = pte_pfn(pte); pte 129 arch/powerpc/platforms/cell/iommu.c static void invalidate_tce_cache(struct cbe_iommu *iommu, unsigned long *pte, pte 142 arch/powerpc/platforms/cell/iommu.c | (__pa(pte) & IOC_IOPT_CacheInvd_IOPTE_Mask) pte 150 arch/powerpc/platforms/cell/iommu.c pte += n; pte 204 arch/powerpc/platforms/cell/iommu.c unsigned long *io_pte, pte; pte 211 arch/powerpc/platforms/cell/iommu.c pte = 0; pte 215 arch/powerpc/platforms/cell/iommu.c pte = CBE_IOPTE_PP_R | CBE_IOPTE_M | CBE_IOPTE_SO_RW | pte 223 arch/powerpc/platforms/cell/iommu.c io_pte[i] = pte; pte 1263 arch/powerpc/platforms/pseries/lpar.c static inline unsigned long compute_slot(real_pte_t pte, pte 1272 arch/powerpc/platforms/pseries/lpar.c hidx = __rpte_to_hidx(pte, index); pte 1290 arch/powerpc/platforms/pseries/lpar.c real_pte_t pte; pte 1298 arch/powerpc/platforms/pseries/lpar.c pte = batch->pte[i]; pte 1299 arch/powerpc/platforms/pseries/lpar.c pte_iterate_hashed_subpages(pte, psize, vpn, index, shift) { pte 1319 arch/powerpc/platforms/pseries/lpar.c slot = compute_slot(pte, vpn, index, shift, ssize); pte 1489 arch/powerpc/platforms/pseries/lpar.c real_pte_t pte; pte 1505 arch/powerpc/platforms/pseries/lpar.c pte = batch->pte[i]; pte 1506 arch/powerpc/platforms/pseries/lpar.c pte_iterate_hashed_subpages(pte, psize, vpn, index, shift) { pte 1507 arch/powerpc/platforms/pseries/lpar.c slot = compute_slot(pte, vpn, index, shift, ssize); pte 3103 arch/powerpc/xmon/xmon.c static void format_pte(void *ptep, unsigned long pte) pte 3105 arch/powerpc/xmon/xmon.c pte_t entry = __pte(pte); pte 3107 arch/powerpc/xmon/xmon.c printf("ptep @ 0x%016lx = 0x%016lx\n", (unsigned long)ptep, pte); pte 3108 arch/powerpc/xmon/xmon.c printf("Maps physical address = 0x%016lx\n", pte & PTE_RPN_MASK); pte 68 arch/riscv/include/asm/page.h unsigned long pte; pte 77 arch/riscv/include/asm/page.h #define pte_val(x) ((x).pte) pte 16 arch/riscv/include/asm/pgalloc.h pmd_t *pmd, pte_t *pte) pte 18 arch/riscv/include/asm/pgalloc.h unsigned long pfn = virt_to_pfn(pte); pte 24 arch/riscv/include/asm/pgalloc.h pmd_t *pmd, pgtable_t pte) pte 26 arch/riscv/include/asm/pgalloc.h unsigned long pfn = virt_to_pfn(page_address(pte)); pte 79 arch/riscv/include/asm/pgalloc.h #define __pte_free_tlb(tlb, pte, buf) \ pte 81 arch/riscv/include/asm/pgalloc.h pgtable_pte_page_dtor(pte); \ pte 82 arch/riscv/include/asm/pgalloc.h tlb_remove_page((tlb), pte); \ pte 179 arch/riscv/include/asm/pgtable.h static inline unsigned long pte_pfn(pte_t pte) pte 181 arch/riscv/include/asm/pgtable.h return (pte_val(pte) >> _PAGE_PFN_SHIFT); pte 202 arch/riscv/include/asm/pgtable.h #define pte_unmap(pte) ((void)(pte)) pte 204 arch/riscv/include/asm/pgtable.h static inline int pte_present(pte_t pte) pte 206 arch/riscv/include/asm/pgtable.h return (pte_val(pte) & (_PAGE_PRESENT | _PAGE_PROT_NONE)); pte 209 arch/riscv/include/asm/pgtable.h static inline int pte_none(pte_t pte) pte 211 arch/riscv/include/asm/pgtable.h return (pte_val(pte) == 0); pte 214 arch/riscv/include/asm/pgtable.h static inline int pte_write(pte_t pte) pte 216 arch/riscv/include/asm/pgtable.h return pte_val(pte) & _PAGE_WRITE; pte 219 arch/riscv/include/asm/pgtable.h static inline int pte_exec(pte_t pte) pte 221 arch/riscv/include/asm/pgtable.h return pte_val(pte) & _PAGE_EXEC; pte 224 arch/riscv/include/asm/pgtable.h static inline int pte_huge(pte_t pte) pte 226 arch/riscv/include/asm/pgtable.h return pte_present(pte) pte 227 arch/riscv/include/asm/pgtable.h && (pte_val(pte) & (_PAGE_READ | _PAGE_WRITE | _PAGE_EXEC)); pte 230 arch/riscv/include/asm/pgtable.h static inline int pte_dirty(pte_t pte) pte 232 arch/riscv/include/asm/pgtable.h return pte_val(pte) & _PAGE_DIRTY; pte 235 arch/riscv/include/asm/pgtable.h static inline int pte_young(pte_t pte) pte 237 arch/riscv/include/asm/pgtable.h return pte_val(pte) & _PAGE_ACCESSED; pte 240 arch/riscv/include/asm/pgtable.h static inline int pte_special(pte_t pte) pte 242 arch/riscv/include/asm/pgtable.h return pte_val(pte) & _PAGE_SPECIAL; pte 247 arch/riscv/include/asm/pgtable.h static inline pte_t pte_wrprotect(pte_t pte) pte 249 arch/riscv/include/asm/pgtable.h return __pte(pte_val(pte) & ~(_PAGE_WRITE)); pte 254 arch/riscv/include/asm/pgtable.h static inline pte_t pte_mkwrite(pte_t pte) pte 256 arch/riscv/include/asm/pgtable.h return __pte(pte_val(pte) | _PAGE_WRITE); pte 261 arch/riscv/include/asm/pgtable.h static inline pte_t pte_mkdirty(pte_t pte) pte 263 arch/riscv/include/asm/pgtable.h return __pte(pte_val(pte) | _PAGE_DIRTY); pte 266 arch/riscv/include/asm/pgtable.h static inline pte_t pte_mkclean(pte_t pte) pte 268 arch/riscv/include/asm/pgtable.h return __pte(pte_val(pte) & ~(_PAGE_DIRTY)); pte 271 arch/riscv/include/asm/pgtable.h static inline pte_t pte_mkyoung(pte_t pte) pte 273 arch/riscv/include/asm/pgtable.h return __pte(pte_val(pte) | _PAGE_ACCESSED); pte 276 arch/riscv/include/asm/pgtable.h static inline pte_t pte_mkold(pte_t pte) pte 278 arch/riscv/include/asm/pgtable.h return __pte(pte_val(pte) & ~(_PAGE_ACCESSED)); pte 281 arch/riscv/include/asm/pgtable.h static inline pte_t pte_mkspecial(pte_t pte) pte 283 arch/riscv/include/asm/pgtable.h return __pte(pte_val(pte) | _PAGE_SPECIAL); pte 286 arch/riscv/include/asm/pgtable.h static inline pte_t pte_mkhuge(pte_t pte) pte 288 arch/riscv/include/asm/pgtable.h return pte; pte 292 arch/riscv/include/asm/pgtable.h static inline pte_t pte_modify(pte_t pte, pgprot_t newprot) pte 294 arch/riscv/include/asm/pgtable.h return __pte((pte_val(pte) & _PAGE_CHG_MASK) | pgprot_val(newprot)); pte 331 arch/riscv/include/asm/pgtable.h void flush_icache_pte(pte_t pte); pte 430 arch/riscv/include/asm/pgtable.h #define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val(pte) }) pte 70 arch/riscv/mm/cacheflush.c void flush_icache_pte(pte_t pte) pte 72 arch/riscv/mm/cacheflush.c struct page *page = pte_page(pte); pte 135 arch/s390/include/asm/gmap.h int gmap_shadow_page(struct gmap *sg, unsigned long saddr, pte_t pte); pte 19 arch/s390/include/asm/hugetlb.h pte_t *ptep, pte_t pte); pte 67 arch/s390/include/asm/hugetlb.h pte_t pte, int dirty) pte 69 arch/s390/include/asm/hugetlb.h int changed = !pte_same(huge_ptep_get(ptep), pte); pte 72 arch/s390/include/asm/hugetlb.h set_huge_pte_at(vma->vm_mm, addr, ptep, pte); pte 80 arch/s390/include/asm/hugetlb.h pte_t pte = huge_ptep_get_and_clear(mm, addr, ptep); pte 81 arch/s390/include/asm/hugetlb.h set_huge_pte_at(mm, addr, ptep, pte_wrprotect(pte)); pte 89 arch/s390/include/asm/hugetlb.h static inline int huge_pte_none(pte_t pte) pte 91 arch/s390/include/asm/hugetlb.h return pte_none(pte); pte 94 arch/s390/include/asm/hugetlb.h static inline int huge_pte_write(pte_t pte) pte 96 arch/s390/include/asm/hugetlb.h return pte_write(pte); pte 99 arch/s390/include/asm/hugetlb.h static inline int huge_pte_dirty(pte_t pte) pte 101 arch/s390/include/asm/hugetlb.h return pte_dirty(pte); pte 104 arch/s390/include/asm/hugetlb.h static inline pte_t huge_pte_mkwrite(pte_t pte) pte 106 arch/s390/include/asm/hugetlb.h return pte_mkwrite(pte); pte 109 arch/s390/include/asm/hugetlb.h static inline pte_t huge_pte_mkdirty(pte_t pte) pte 111 arch/s390/include/asm/hugetlb.h return pte_mkdirty(pte); pte 114 arch/s390/include/asm/hugetlb.h static inline pte_t huge_pte_wrprotect(pte_t pte) pte 116 arch/s390/include/asm/hugetlb.h return pte_wrprotect(pte); pte 119 arch/s390/include/asm/hugetlb.h static inline pte_t huge_pte_modify(pte_t pte, pgprot_t newprot) pte 121 arch/s390/include/asm/hugetlb.h return pte_modify(pte, newprot); pte 81 arch/s390/include/asm/page.h typedef struct { unsigned long pte; } pte_t; pte 90 arch/s390/include/asm/page.h #define pte_val(x) ((x).pte) pte 141 arch/s390/include/asm/pgalloc.h pmd_t *pmd, pgtable_t pte) pte 143 arch/s390/include/asm/pgalloc.h pmd_val(*pmd) = _SEGMENT_ENTRY + __pa(pte); pte 146 arch/s390/include/asm/pgalloc.h #define pmd_populate_kernel(mm, pmd, pte) pmd_populate(mm, pmd, pte) pte 157 arch/s390/include/asm/pgalloc.h #define pte_free_kernel(mm, pte) page_table_free(mm, (unsigned long *) pte) pte 158 arch/s390/include/asm/pgalloc.h #define pte_free(mm, pte) page_table_free(mm, (unsigned long *) pte) pte 781 arch/s390/include/asm/pgtable.h static inline int pte_present(pte_t pte) pte 784 arch/s390/include/asm/pgtable.h return (pte_val(pte) & _PAGE_PRESENT) != 0; pte 787 arch/s390/include/asm/pgtable.h static inline int pte_none(pte_t pte) pte 790 arch/s390/include/asm/pgtable.h return pte_val(pte) == _PAGE_INVALID; pte 793 arch/s390/include/asm/pgtable.h static inline int pte_swap(pte_t pte) pte 796 arch/s390/include/asm/pgtable.h return (pte_val(pte) & (_PAGE_PROTECT | _PAGE_PRESENT)) pte 800 arch/s390/include/asm/pgtable.h static inline int pte_special(pte_t pte) pte 802 arch/s390/include/asm/pgtable.h return (pte_val(pte) & _PAGE_SPECIAL); pte 812 arch/s390/include/asm/pgtable.h static inline int pte_protnone(pte_t pte) pte 814 arch/s390/include/asm/pgtable.h return pte_present(pte) && !(pte_val(pte) & _PAGE_READ); pte 824 arch/s390/include/asm/pgtable.h static inline int pte_soft_dirty(pte_t pte) pte 826 arch/s390/include/asm/pgtable.h return pte_val(pte) & _PAGE_SOFT_DIRTY; pte 830 arch/s390/include/asm/pgtable.h static inline pte_t pte_mksoft_dirty(pte_t pte) pte 832 arch/s390/include/asm/pgtable.h pte_val(pte) |= _PAGE_SOFT_DIRTY; pte 833 arch/s390/include/asm/pgtable.h return pte; pte 837 arch/s390/include/asm/pgtable.h static inline pte_t pte_clear_soft_dirty(pte_t pte) pte 839 arch/s390/include/asm/pgtable.h pte_val(pte) &= ~_PAGE_SOFT_DIRTY; pte 840 arch/s390/include/asm/pgtable.h return pte; pte 865 arch/s390/include/asm/pgtable.h static inline int pte_write(pte_t pte) pte 867 arch/s390/include/asm/pgtable.h return (pte_val(pte) & _PAGE_WRITE) != 0; pte 870 arch/s390/include/asm/pgtable.h static inline int pte_dirty(pte_t pte) pte 872 arch/s390/include/asm/pgtable.h return (pte_val(pte) & _PAGE_DIRTY) != 0; pte 875 arch/s390/include/asm/pgtable.h static inline int pte_young(pte_t pte) pte 877 arch/s390/include/asm/pgtable.h return (pte_val(pte) & _PAGE_YOUNG) != 0; pte 881 arch/s390/include/asm/pgtable.h static inline int pte_unused(pte_t pte) pte 883 arch/s390/include/asm/pgtable.h return pte_val(pte) & _PAGE_UNUSED; pte 922 arch/s390/include/asm/pgtable.h static inline pte_t pte_modify(pte_t pte, pgprot_t newprot) pte 924 arch/s390/include/asm/pgtable.h pte_val(pte) &= _PAGE_CHG_MASK; pte 925 arch/s390/include/asm/pgtable.h pte_val(pte) |= pgprot_val(newprot); pte 930 arch/s390/include/asm/pgtable.h if ((pte_val(pte) & _PAGE_YOUNG) && (pte_val(pte) & _PAGE_READ)) pte 931 arch/s390/include/asm/pgtable.h pte_val(pte) &= ~_PAGE_INVALID; pte 936 arch/s390/include/asm/pgtable.h if ((pte_val(pte) & _PAGE_DIRTY) && (pte_val(pte) & _PAGE_WRITE)) pte 937 arch/s390/include/asm/pgtable.h pte_val(pte) &= ~_PAGE_PROTECT; pte 938 arch/s390/include/asm/pgtable.h return pte; pte 941 arch/s390/include/asm/pgtable.h static inline pte_t pte_wrprotect(pte_t pte) pte 943 arch/s390/include/asm/pgtable.h pte_val(pte) &= ~_PAGE_WRITE; pte 944 arch/s390/include/asm/pgtable.h pte_val(pte) |= _PAGE_PROTECT; pte 945 arch/s390/include/asm/pgtable.h return pte; pte 948 arch/s390/include/asm/pgtable.h static inline pte_t pte_mkwrite(pte_t pte) pte 950 arch/s390/include/asm/pgtable.h pte_val(pte) |= _PAGE_WRITE; pte 951 arch/s390/include/asm/pgtable.h if (pte_val(pte) & _PAGE_DIRTY) pte 952 arch/s390/include/asm/pgtable.h pte_val(pte) &= ~_PAGE_PROTECT; pte 953 arch/s390/include/asm/pgtable.h return pte; pte 956 arch/s390/include/asm/pgtable.h static inline pte_t pte_mkclean(pte_t pte) pte 958 arch/s390/include/asm/pgtable.h pte_val(pte) &= ~_PAGE_DIRTY; pte 959 arch/s390/include/asm/pgtable.h pte_val(pte) |= _PAGE_PROTECT; pte 960 arch/s390/include/asm/pgtable.h return pte; pte 963 arch/s390/include/asm/pgtable.h static inline pte_t pte_mkdirty(pte_t pte) pte 965 arch/s390/include/asm/pgtable.h pte_val(pte) |= _PAGE_DIRTY | _PAGE_SOFT_DIRTY; pte 966 arch/s390/include/asm/pgtable.h if (pte_val(pte) & _PAGE_WRITE) pte 967 arch/s390/include/asm/pgtable.h pte_val(pte) &= ~_PAGE_PROTECT; pte 968 arch/s390/include/asm/pgtable.h return pte; pte 971 arch/s390/include/asm/pgtable.h static inline pte_t pte_mkold(pte_t pte) pte 973 arch/s390/include/asm/pgtable.h pte_val(pte) &= ~_PAGE_YOUNG; pte 974 arch/s390/include/asm/pgtable.h pte_val(pte) |= _PAGE_INVALID; pte 975 arch/s390/include/asm/pgtable.h return pte; pte 978 arch/s390/include/asm/pgtable.h static inline pte_t pte_mkyoung(pte_t pte) pte 980 arch/s390/include/asm/pgtable.h pte_val(pte) |= _PAGE_YOUNG; pte 981 arch/s390/include/asm/pgtable.h if (pte_val(pte) & _PAGE_READ) pte 982 arch/s390/include/asm/pgtable.h pte_val(pte) &= ~_PAGE_INVALID; pte 983 arch/s390/include/asm/pgtable.h return pte; pte 986 arch/s390/include/asm/pgtable.h static inline pte_t pte_mkspecial(pte_t pte) pte 988 arch/s390/include/asm/pgtable.h pte_val(pte) |= _PAGE_SPECIAL; pte 989 arch/s390/include/asm/pgtable.h return pte; pte 993 arch/s390/include/asm/pgtable.h static inline pte_t pte_mkhuge(pte_t pte) pte 995 arch/s390/include/asm/pgtable.h pte_val(pte) |= _PAGE_LARGE; pte 996 arch/s390/include/asm/pgtable.h return pte; pte 1063 arch/s390/include/asm/pgtable.h pte_t pte = *ptep; pte 1065 arch/s390/include/asm/pgtable.h pte = ptep_xchg_direct(vma->vm_mm, addr, ptep, pte_mkold(pte)); pte 1066 arch/s390/include/asm/pgtable.h return pte_young(pte); pte 1108 arch/s390/include/asm/pgtable.h pte_t pte = *ptep; pte 1110 arch/s390/include/asm/pgtable.h return pte; pte 1119 arch/s390/include/asm/pgtable.h pte_t pte = *ptep; pte 1121 arch/s390/include/asm/pgtable.h if (pte_write(pte)) pte 1122 arch/s390/include/asm/pgtable.h ptep_xchg_lazy(mm, addr, ptep, pte_wrprotect(pte)); pte 1150 arch/s390/include/asm/pgtable.h pte_t *sptep, pte_t *tptep, pte_t pte); pte 1279 arch/s390/include/asm/pgtable.h static inline void pte_unmap(pte_t *pte) { } pte 1654 arch/s390/include/asm/pgtable.h pte_t pte; pte 1656 arch/s390/include/asm/pgtable.h pte_val(pte) = _PAGE_INVALID | _PAGE_PROTECT; pte 1657 arch/s390/include/asm/pgtable.h pte_val(pte) |= (offset & __SWP_OFFSET_MASK) << __SWP_OFFSET_SHIFT; pte 1658 arch/s390/include/asm/pgtable.h pte_val(pte) |= (type & __SWP_TYPE_MASK) << __SWP_TYPE_SHIFT; pte 1659 arch/s390/include/asm/pgtable.h return pte; pte 1677 arch/s390/include/asm/pgtable.h #define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val(pte) }) pte 64 arch/s390/include/asm/tlb.h static inline void pte_free_tlb(struct mmu_gather *tlb, pgtable_t pte, pte 76 arch/s390/include/asm/tlb.h page_table_free_rcu(tlb, (unsigned long *) pte, address); pte 620 arch/s390/kvm/gaccess.c union page_table_entry pte; pte 752 arch/s390/kvm/gaccess.c if (deref_table(vcpu->kvm, ptr, &pte.val)) pte 754 arch/s390/kvm/gaccess.c if (pte.i) pte 756 arch/s390/kvm/gaccess.c if (pte.z) pte 758 arch/s390/kvm/gaccess.c dat_protection |= pte.p; pte 759 arch/s390/kvm/gaccess.c iep_protection = pte.iep; pte 760 arch/s390/kvm/gaccess.c raddr.pfra = pte.pfra; pte 1168 arch/s390/kvm/gaccess.c union page_table_entry pte; pte 1188 arch/s390/kvm/gaccess.c pte.val = pgt + vaddr.px * PAGE_SIZE; pte 1192 arch/s390/kvm/gaccess.c rc = gmap_read_table(sg->parent, pgt + vaddr.px * 8, &pte.val); pte 1193 arch/s390/kvm/gaccess.c if (!rc && pte.i) pte 1195 arch/s390/kvm/gaccess.c if (!rc && pte.z) pte 1198 arch/s390/kvm/gaccess.c pte.p |= dat_protection; pte 1200 arch/s390/kvm/gaccess.c rc = gmap_shadow_page(sg, saddr, __pte(pte.val)); pte 137 arch/s390/mm/dump_pagetables.c pte_t *pte; pte 142 arch/s390/mm/dump_pagetables.c pte = pte_offset_kernel(pmd, addr); pte 143 arch/s390/mm/dump_pagetables.c prot = pte_val(*pte) & pte 1133 arch/s390/mm/gmap.c pte_t *ptep, pte; pte 1143 arch/s390/mm/gmap.c pte = *ptep; pte 1144 arch/s390/mm/gmap.c if (pte_present(pte) && (pte_val(pte) & _PAGE_READ)) { pte 1145 arch/s390/mm/gmap.c address = pte_val(pte) & PAGE_MASK; pte 2105 arch/s390/mm/gmap.c int gmap_shadow_page(struct gmap *sg, unsigned long saddr, pte_t pte) pte 2117 arch/s390/mm/gmap.c prot = (pte_val(pte) & _PAGE_PROTECT) ? PROT_READ : PROT_WRITE; pte 2125 arch/s390/mm/gmap.c paddr = pte_val(pte) & PAGE_MASK; pte 2146 arch/s390/mm/gmap.c rc = ptep_shadow_pte(sg->mm, saddr, sptep, tptep, pte); pte 2236 arch/s390/mm/gmap.c pte_t *pte, unsigned long bits) pte 2242 arch/s390/mm/gmap.c offset = ((unsigned long) pte) & (255 * sizeof(pte_t)); pte 2560 arch/s390/mm/gmap.c static int __s390_enable_skey_pte(pte_t *pte, unsigned long addr, pte 2564 arch/s390/mm/gmap.c ptep_zap_key(walk->mm, addr, pte); pte 2568 arch/s390/mm/gmap.c static int __s390_enable_skey_hugetlb(pte_t *pte, unsigned long addr, pte 2572 arch/s390/mm/gmap.c pmd_t *pmd = (pmd_t *)pte; pte 2630 arch/s390/mm/gmap.c static int __s390_reset_cmma(pte_t *pte, unsigned long addr, pte 2633 arch/s390/mm/gmap.c ptep_zap_unused(walk->mm, addr, pte, 1); pte 24 arch/s390/mm/hugetlbpage.c static inline unsigned long __pte_to_rste(pte_t pte) pte 48 arch/s390/mm/hugetlbpage.c if (pte_present(pte)) { pte 49 arch/s390/mm/hugetlbpage.c rste = pte_val(pte) & PAGE_MASK; pte 50 arch/s390/mm/hugetlbpage.c rste |= move_set_bit(pte_val(pte), _PAGE_READ, pte 52 arch/s390/mm/hugetlbpage.c rste |= move_set_bit(pte_val(pte), _PAGE_WRITE, pte 54 arch/s390/mm/hugetlbpage.c rste |= move_set_bit(pte_val(pte), _PAGE_INVALID, pte 56 arch/s390/mm/hugetlbpage.c rste |= move_set_bit(pte_val(pte), _PAGE_PROTECT, pte 58 arch/s390/mm/hugetlbpage.c rste |= move_set_bit(pte_val(pte), _PAGE_DIRTY, pte 60 arch/s390/mm/hugetlbpage.c rste |= move_set_bit(pte_val(pte), _PAGE_YOUNG, pte 63 arch/s390/mm/hugetlbpage.c rste |= move_set_bit(pte_val(pte), _PAGE_SOFT_DIRTY, pte 66 arch/s390/mm/hugetlbpage.c rste |= move_set_bit(pte_val(pte), _PAGE_NOEXEC, pte 76 arch/s390/mm/hugetlbpage.c pte_t pte; pte 104 arch/s390/mm/hugetlbpage.c pte_val(pte) = rste & _SEGMENT_ENTRY_ORIGIN_LARGE; pte 105 arch/s390/mm/hugetlbpage.c pte_val(pte) |= _PAGE_LARGE | _PAGE_PRESENT; pte 106 arch/s390/mm/hugetlbpage.c pte_val(pte) |= move_set_bit(rste, _SEGMENT_ENTRY_READ, pte 108 arch/s390/mm/hugetlbpage.c pte_val(pte) |= move_set_bit(rste, _SEGMENT_ENTRY_WRITE, pte 110 arch/s390/mm/hugetlbpage.c pte_val(pte) |= move_set_bit(rste, _SEGMENT_ENTRY_INVALID, pte 112 arch/s390/mm/hugetlbpage.c pte_val(pte) |= move_set_bit(rste, _SEGMENT_ENTRY_PROTECT, pte 114 arch/s390/mm/hugetlbpage.c pte_val(pte) |= move_set_bit(rste, _SEGMENT_ENTRY_DIRTY, pte 116 arch/s390/mm/hugetlbpage.c pte_val(pte) |= move_set_bit(rste, _SEGMENT_ENTRY_YOUNG, pte 119 arch/s390/mm/hugetlbpage.c pte_val(pte) |= move_set_bit(rste, _SEGMENT_ENTRY_SOFT_DIRTY, pte 122 arch/s390/mm/hugetlbpage.c pte_val(pte) |= move_set_bit(rste, _SEGMENT_ENTRY_NOEXEC, pte 125 arch/s390/mm/hugetlbpage.c pte_val(pte) = _PAGE_INVALID; pte 126 arch/s390/mm/hugetlbpage.c return pte; pte 153 arch/s390/mm/hugetlbpage.c pte_t *ptep, pte_t pte) pte 157 arch/s390/mm/hugetlbpage.c rste = __pte_to_rste(pte); pte 163 arch/s390/mm/hugetlbpage.c if (likely(pte_present(pte))) pte 166 arch/s390/mm/hugetlbpage.c } else if (likely(pte_present(pte))) pte 181 arch/s390/mm/hugetlbpage.c pte_t pte = huge_ptep_get(ptep); pte 189 arch/s390/mm/hugetlbpage.c return pte; pte 67 arch/s390/mm/kasan_init.c pte_t *pte; pte 73 arch/s390/mm/kasan_init.c pte = pte_leftover + _PAGE_TABLE_SIZE; pte 75 arch/s390/mm/kasan_init.c pte = pte_leftover; pte 78 arch/s390/mm/kasan_init.c memset64((u64 *)pte, _PAGE_INVALID, PTRS_PER_PTE); pte 79 arch/s390/mm/kasan_init.c return pte; pte 322 arch/s390/mm/pageattr.c static void ipte_range(pte_t *pte, unsigned long address, int nr) pte 327 arch/s390/mm/pageattr.c __ptep_ipte_range(address, nr - 1, pte, IPTE_GLOBAL); pte 331 arch/s390/mm/pageattr.c __ptep_ipte(address, pte, 0, 0, IPTE_GLOBAL); pte 333 arch/s390/mm/pageattr.c pte++; pte 345 arch/s390/mm/pageattr.c pte_t *pte; pte 353 arch/s390/mm/pageattr.c pte = pte_offset_kernel(pmd, address); pte 354 arch/s390/mm/pageattr.c nr = (unsigned long)pte >> ilog2(sizeof(long)); pte 359 arch/s390/mm/pageattr.c pte_val(*pte) &= ~_PAGE_INVALID; pte 361 arch/s390/mm/pageattr.c pte++; pte 364 arch/s390/mm/pageattr.c ipte_range(pte, address, nr); pte 394 arch/s390/mm/pgalloc.c unsigned long *pte, next; pte 398 arch/s390/mm/pgalloc.c pte = (unsigned long *) origin; pte 399 arch/s390/mm/pgalloc.c pte += (addr & _PAGE_INDEX) >> _PAGE_SHIFT; pte 402 arch/s390/mm/pgalloc.c *pte = base_lra(addr); pte 403 arch/s390/mm/pgalloc.c } while (pte++, addr = next, addr < end); pte 155 arch/s390/mm/pgtable.c static inline pgste_t pgste_update_all(pte_t pte, pgste_t pgste, pte 161 arch/s390/mm/pgtable.c if (!mm_uses_skeys(mm) || pte_val(pte) & _PAGE_INVALID) pte 163 arch/s390/mm/pgtable.c address = pte_val(pte) & PAGE_MASK; pte 324 arch/s390/mm/pgtable.c pte_t *ptep, pte_t old_pte, pte_t pte) pte 330 arch/s390/mm/pgtable.c pte_val(pte) &= ~_PAGE_NOEXEC; pte 333 arch/s390/mm/pgtable.c pgste_set_key(ptep, pgste, pte, mm); pte 334 arch/s390/mm/pgtable.c pgste = pgste_set_pte(ptep, pgste, pte); pte 337 arch/s390/mm/pgtable.c *ptep = pte; pte 631 arch/s390/mm/pgtable.c pte_t *sptep, pte_t *tptep, pte_t pte) pte 643 arch/s390/mm/pgtable.c !(pte_val(pte) & _PAGE_PROTECT))) { pte 647 arch/s390/mm/pgtable.c (pte_val(pte) & _PAGE_PROTECT); pte 688 arch/s390/mm/pgtable.c pte_t pte; pte 694 arch/s390/mm/pgtable.c pte = *ptep; pte 695 arch/s390/mm/pgtable.c if (!reset && pte_swap(pte) && pte 698 arch/s390/mm/pgtable.c ptep_zap_swap_entry(mm, pte_to_swp_entry(pte)); pte 731 arch/s390/mm/pgtable.c pte_t pte; pte 738 arch/s390/mm/pgtable.c pte = *ptep; pte 739 arch/s390/mm/pgtable.c if (dirty && (pte_val(pte) & _PAGE_PRESENT)) { pte 743 arch/s390/mm/pgtable.c if (MACHINE_HAS_ESOP || !(pte_val(pte) & _PAGE_WRITE)) pte 744 arch/s390/mm/pgtable.c pte_val(pte) |= _PAGE_PROTECT; pte 746 arch/s390/mm/pgtable.c pte_val(pte) |= _PAGE_INVALID; pte 747 arch/s390/mm/pgtable.c *ptep = pte; pte 54 arch/s390/mm/vmem.c pte_t *pte; pte 57 arch/s390/mm/vmem.c pte = (pte_t *) page_table_alloc(&init_mm); pte 59 arch/s390/mm/vmem.c pte = (pte_t *) memblock_phys_alloc(size, size); pte 60 arch/s390/mm/vmem.c if (!pte) pte 62 arch/s390/mm/vmem.c memset64((u64 *)pte, _PAGE_INVALID, PTRS_PER_PTE); pte 63 arch/s390/mm/vmem.c return pte; pte 18 arch/sh/include/asm/pgalloc.h pte_t *pte) pte 20 arch/sh/include/asm/pgalloc.h set_pmd(pmd, __pmd((unsigned long)pte)); pte 24 arch/sh/include/asm/pgalloc.h pgtable_t pte) pte 26 arch/sh/include/asm/pgalloc.h set_pmd(pmd, __pmd((unsigned long)page_address(pte))); pte 30 arch/sh/include/asm/pgalloc.h #define __pte_free_tlb(tlb,pte,addr) \ pte 32 arch/sh/include/asm/pgalloc.h pgtable_pte_page_dtor(pte); \ pte 33 arch/sh/include/asm/pgalloc.h tlb_remove_page((tlb), (pte)); \ pte 130 arch/sh/include/asm/pgtable.h unsigned long address, pte_t pte); pte 132 arch/sh/include/asm/pgtable.h unsigned long address, pte_t pte); pte 137 arch/sh/include/asm/pgtable.h pte_t pte = *ptep; pte 138 arch/sh/include/asm/pgtable.h __update_cache(vma, address, pte); pte 139 arch/sh/include/asm/pgtable.h __update_tlb(vma, address, pte); pte 147 arch/sh/include/asm/pgtable.h static inline bool __pte_access_permitted(pte_t pte, u64 prot) pte 149 arch/sh/include/asm/pgtable.h return (pte_val(pte) & (prot | _PAGE_SPECIAL)) == prot; pte 153 arch/sh/include/asm/pgtable.h static inline bool pte_access_permitted(pte_t pte, bool write) pte 160 arch/sh/include/asm/pgtable.h return __pte_access_permitted(pte, prot); pte 163 arch/sh/include/asm/pgtable.h static inline bool pte_access_permitted(pte_t pte, bool write) pte 169 arch/sh/include/asm/pgtable.h return __pte_access_permitted(pte, prot); pte 172 arch/sh/include/asm/pgtable.h static inline bool pte_access_permitted(pte_t pte, bool write) pte 178 arch/sh/include/asm/pgtable.h return __pte_access_permitted(pte, prot); pte 300 arch/sh/include/asm/pgtable_32.h static inline void set_pte(pte_t *ptep, pte_t pte) pte 302 arch/sh/include/asm/pgtable_32.h ptep->pte_high = pte.pte_high; pte 304 arch/sh/include/asm/pgtable_32.h ptep->pte_low = pte.pte_low; pte 340 arch/sh/include/asm/pgtable_32.h #define pte_not_present(pte) (!((pte).pte_low & _PAGE_PRESENT)) pte 341 arch/sh/include/asm/pgtable_32.h #define pte_dirty(pte) ((pte).pte_low & _PAGE_DIRTY) pte 342 arch/sh/include/asm/pgtable_32.h #define pte_young(pte) ((pte).pte_low & _PAGE_ACCESSED) pte 343 arch/sh/include/asm/pgtable_32.h #define pte_special(pte) ((pte).pte_low & _PAGE_SPECIAL) pte 346 arch/sh/include/asm/pgtable_32.h #define pte_write(pte) \ pte 347 arch/sh/include/asm/pgtable_32.h ((pte).pte_high & (_PAGE_EXT_USER_WRITE | _PAGE_EXT_KERN_WRITE)) pte 349 arch/sh/include/asm/pgtable_32.h #define pte_write(pte) ((pte).pte_low & _PAGE_RW) pte 353 arch/sh/include/asm/pgtable_32.h static inline pte_t pte_##fn(pte_t pte) { pte.pte_##h op; return pte; } pte 392 arch/sh/include/asm/pgtable_32.h static inline pte_t pte_modify(pte_t pte, pgprot_t newprot) pte 394 arch/sh/include/asm/pgtable_32.h pte.pte_low &= _PAGE_CHG_MASK; pte 395 arch/sh/include/asm/pgtable_32.h pte.pte_low |= pgprot_val(newprot); pte 398 arch/sh/include/asm/pgtable_32.h pte.pte_high |= pgprot_val(newprot) >> 32; pte 401 arch/sh/include/asm/pgtable_32.h return pte; pte 425 arch/sh/include/asm/pgtable_32.h #define pte_unmap(pte) do { } while (0) pte 464 arch/sh/include/asm/pgtable_32.h #define __pte_to_swp_entry(pte) ((swp_entry_t){ (pte).pte_high }) pte 472 arch/sh/include/asm/pgtable_32.h #define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val(pte) >> 1 }) pte 84 arch/sh/include/asm/pgtable_64.h #define pte_unmap(pte) do { } while (0) pte 257 arch/sh/include/asm/pgtable_64.h static inline int pte_dirty(pte_t pte) { return pte_val(pte) & _PAGE_DIRTY; } pte 258 arch/sh/include/asm/pgtable_64.h static inline int pte_young(pte_t pte) { return pte_val(pte) & _PAGE_ACCESSED; } pte 259 arch/sh/include/asm/pgtable_64.h static inline int pte_write(pte_t pte) { return pte_val(pte) & _PAGE_WRITE; } pte 260 arch/sh/include/asm/pgtable_64.h static inline int pte_special(pte_t pte){ return pte_val(pte) & _PAGE_SPECIAL; } pte 262 arch/sh/include/asm/pgtable_64.h static inline pte_t pte_wrprotect(pte_t pte) { set_pte(&pte, __pte(pte_val(pte) & ~_PAGE_WRITE)); return pte; } pte 263 arch/sh/include/asm/pgtable_64.h static inline pte_t pte_mkclean(pte_t pte) { set_pte(&pte, __pte(pte_val(pte) & ~_PAGE_DIRTY)); return pte; } pte 264 arch/sh/include/asm/pgtable_64.h static inline pte_t pte_mkold(pte_t pte) { set_pte(&pte, __pte(pte_val(pte) & ~_PAGE_ACCESSED)); return pte; } pte 265 arch/sh/include/asm/pgtable_64.h static inline pte_t pte_mkwrite(pte_t pte) { set_pte(&pte, __pte(pte_val(pte) | _PAGE_WRITE)); return pte; } pte 266 arch/sh/include/asm/pgtable_64.h static inline pte_t pte_mkdirty(pte_t pte) { set_pte(&pte, __pte(pte_val(pte) | _PAGE_DIRTY)); return pte; } pte 267 arch/sh/include/asm/pgtable_64.h static inline pte_t pte_mkyoung(pte_t pte) { set_pte(&pte, __pte(pte_val(pte) | _PAGE_ACCESSED)); return pte; } pte 268 arch/sh/include/asm/pgtable_64.h static inline pte_t pte_mkhuge(pte_t pte) { set_pte(&pte, __pte(pte_val(pte) | _PAGE_SZHUGE)); return pte; } pte 269 arch/sh/include/asm/pgtable_64.h static inline pte_t pte_mkspecial(pte_t pte) { set_pte(&pte, __pte(pte_val(pte) | _PAGE_SPECIAL)); return pte; } pte 292 arch/sh/include/asm/pgtable_64.h static inline pte_t pte_modify(pte_t pte, pgprot_t newprot) pte 293 arch/sh/include/asm/pgtable_64.h { set_pte(&pte, __pte((pte_val(pte) & _PAGE_CHG_MASK) | pgprot_val(newprot))); return pte; } pte 299 arch/sh/include/asm/pgtable_64.h #define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val(pte) }) pte 22 arch/sh/include/asm/tlb.h unsigned long addr, pte_t pte) pte 19 arch/sh/include/cpu-sh5/cpu/mmu_context.h extern void __do_tlb_refill(unsigned long address, unsigned long long is_text_not_data, pte_t *pte); pte 214 arch/sh/mm/cache-sh4.c pte_t *pte; pte 229 arch/sh/mm/cache-sh4.c pte = pte_offset_kernel(pmd, address); pte 232 arch/sh/mm/cache-sh4.c if (!(pte_val(*pte) & _PAGE_PRESENT)) pte 388 arch/sh/mm/cache-sh5.c pte_t *pte; pte 408 arch/sh/mm/cache-sh5.c pte = pte_offset_map_lock(mm, pmd, addr, &ptl); pte 410 arch/sh/mm/cache-sh5.c entry = *pte; pte 415 arch/sh/mm/cache-sh5.c } while (pte++, addr += PAGE_SIZE, addr != end); pte 416 arch/sh/mm/cache-sh5.c pte_unmap_unlock(pte - 1, ptl); pte 137 arch/sh/mm/cache.c unsigned long address, pte_t pte) pte 140 arch/sh/mm/cache.c unsigned long pfn = pte_pfn(pte); pte 58 arch/sh/mm/fault.c pte_t *pte; pte 98 arch/sh/mm/fault.c pte = pte_offset_kernel(pmd, addr); pte 99 arch/sh/mm/fault.c printk(", *pte=%0*Lx", (u32)(sizeof(*pte) * 2), pte 100 arch/sh/mm/fault.c (u64)pte_val(*pte)); pte 31 arch/sh/mm/hugetlbpage.c pte_t *pte = NULL; pte 39 arch/sh/mm/hugetlbpage.c pte = pte_alloc_map(mm, pmd, addr); pte 43 arch/sh/mm/hugetlbpage.c return pte; pte 52 arch/sh/mm/hugetlbpage.c pte_t *pte = NULL; pte 60 arch/sh/mm/hugetlbpage.c pte = pte_offset_map(pmd, addr); pte 64 arch/sh/mm/hugetlbpage.c return pte; pte 74 arch/sh/mm/init.c pte_t *pte; pte 76 arch/sh/mm/init.c pte = __get_pte_phys(addr); pte 77 arch/sh/mm/init.c if (!pte_none(*pte)) { pte 78 arch/sh/mm/init.c pte_ERROR(*pte); pte 82 arch/sh/mm/init.c set_pte(pte, pfn_pte(phys >> PAGE_SHIFT, prot)); pte 86 arch/sh/mm/init.c tlb_wire_entry(NULL, addr, *pte); pte 91 arch/sh/mm/init.c pte_t *pte; pte 93 arch/sh/mm/init.c pte = __get_pte_phys(addr); pte 98 arch/sh/mm/init.c set_pte(pte, pfn_pte(0, __pgprot(0))); pte 145 arch/sh/mm/init.c pte_t *pte; pte 147 arch/sh/mm/init.c pte = memblock_alloc(PAGE_SIZE, PAGE_SIZE); pte 148 arch/sh/mm/init.c if (!pte) pte 151 arch/sh/mm/init.c pmd_populate_kernel(&init_mm, pmd, pte); pte 152 arch/sh/mm/init.c BUG_ON(pte != pte_offset_kernel(pmd, 0)); pte 158 arch/sh/mm/init.c static pte_t * __init page_table_kmap_check(pte_t *pte, pmd_t *pmd, pte 161 arch/sh/mm/init.c return pte; pte 170 arch/sh/mm/init.c pte_t *pte = NULL; pte 188 arch/sh/mm/init.c pte = page_table_kmap_check(one_page_table_init(pmd), pte 189 arch/sh/mm/init.c pmd, vaddr, pte); pte 73 arch/sh/mm/nommu.c void __update_tlb(struct vm_area_struct *vma, unsigned long address, pte_t pte) pte 18 arch/sh/mm/tlb-pteaex.c void __update_tlb(struct vm_area_struct *vma, unsigned long address, pte_t pte) pte 37 arch/sh/mm/tlb-pteaex.c pteval = pte.pte_low; pte 47 arch/sh/mm/tlb-pteaex.c __raw_writel(pte.pte_high, MMU_PTEA); pte 28 arch/sh/mm/tlb-sh3.c void __update_tlb(struct vm_area_struct *vma, unsigned long address, pte_t pte) pte 44 arch/sh/mm/tlb-sh3.c pteval = pte_val(pte); pte 16 arch/sh/mm/tlb-sh4.c void __update_tlb(struct vm_area_struct *vma, unsigned long address, pte_t pte) pte 32 arch/sh/mm/tlb-sh4.c pteval = pte.pte_low; pte 42 arch/sh/mm/tlb-sh4.c __raw_writel(pte.pte_high, MMU_PTEA); pte 150 arch/sh/mm/tlb-sh5.c void tlb_wire_entry(struct vm_area_struct *vma, unsigned long addr, pte_t pte) pte 162 arch/sh/mm/tlb-sh5.c paddr = pte_val(pte) & _PAGE_FLAGS_HARDWARE_MASK; pte 186 arch/sh/mm/tlb-sh5.c void __update_tlb(struct vm_area_struct *vma, unsigned long address, pte_t pte) pte 195 arch/sh/mm/tlb-sh5.c ptel = pte.pte_low; pte 20 arch/sh/mm/tlb-urb.c void tlb_wire_entry(struct vm_area_struct *vma, unsigned long addr, pte_t pte) pte 47 arch/sh/mm/tlb-urb.c __update_tlb(vma, addr, pte); pte 28 arch/sh/mm/tlbex_32.c pte_t *pte; pte 51 arch/sh/mm/tlbex_32.c pte = pte_offset_kernel(pmd, address); pte 52 arch/sh/mm/tlbex_32.c entry = *pte; pte 62 arch/sh/mm/tlbex_32.c set_pte(pte, entry); pte 75 arch/sh/mm/tlbex_32.c update_mmu_cache(NULL, address, pte); pte 49 arch/sh/mm/tlbex_64.c pte_t *pte; pte 69 arch/sh/mm/tlbex_64.c pte = pte_offset_kernel(pmd, address); pte 70 arch/sh/mm/tlbex_64.c entry = *pte; pte 83 arch/sh/mm/tlbex_64.c update_mmu_cache(NULL, address, pte); pte 17 arch/sparc/include/asm/hugetlb.h pte_t *ptep, pte_t pte); pte 46 arch/sparc/include/asm/hugetlb.h pte_t pte, int dirty) pte 48 arch/sparc/include/asm/hugetlb.h int changed = !pte_same(*ptep, pte); pte 50 arch/sparc/include/asm/hugetlb.h set_huge_pte_at(vma->vm_mm, addr, ptep, pte); pte 68 arch/sparc/include/asm/mmu_64.h unsigned long pte; pte 71 arch/sparc/include/asm/mmu_64.h void __tsb_insert(unsigned long ent, unsigned long tag, unsigned long pte); pte 55 arch/sparc/include/asm/page_32.h typedef struct { unsigned long pte; } pte_t; pte 63 arch/sparc/include/asm/page_32.h #define pte_val(x) ((x).pte) pte 69 arch/sparc/include/asm/page_64.h typedef struct { unsigned long pte; } pte_t; pte 76 arch/sparc/include/asm/page_64.h #define pte_val(x) ((x).pte) pte 67 arch/sparc/include/asm/pgalloc_32.h static inline void free_pte_fast(pte_t *pte) pte 69 arch/sparc/include/asm/pgalloc_32.h srmmu_free_nocache(pte, PTE_SIZE); pte 72 arch/sparc/include/asm/pgalloc_32.h #define pte_free_kernel(mm, pte) free_pte_fast(pte) pte 74 arch/sparc/include/asm/pgalloc_32.h void pte_free(struct mm_struct * mm, pgtable_t pte); pte 75 arch/sparc/include/asm/pgalloc_32.h #define __pte_free_tlb(tlb, pte, addr) pte_free((tlb)->mm, pte) pte 65 arch/sparc/include/asm/pgalloc_64.h void pte_free_kernel(struct mm_struct *mm, pte_t *pte); pte 103 arch/sparc/include/asm/pgalloc_64.h static inline void __pte_free_tlb(struct mmu_gather *tlb, pte_t *pte, pte 106 arch/sparc/include/asm/pgalloc_64.h pgtable_free_tlb(tlb, pte, true); pte 145 arch/sparc/include/asm/pgtable_32.h static inline int pte_present(pte_t pte) pte 147 arch/sparc/include/asm/pgtable_32.h return ((pte_val(pte) & SRMMU_ET_MASK) == SRMMU_ET_PTE); pte 150 arch/sparc/include/asm/pgtable_32.h static inline int pte_none(pte_t pte) pte 152 arch/sparc/include/asm/pgtable_32.h return !pte_val(pte); pte 211 arch/sparc/include/asm/pgtable_32.h static inline int pte_write(pte_t pte) pte 213 arch/sparc/include/asm/pgtable_32.h return pte_val(pte) & SRMMU_WRITE; pte 216 arch/sparc/include/asm/pgtable_32.h static inline int pte_dirty(pte_t pte) pte 218 arch/sparc/include/asm/pgtable_32.h return pte_val(pte) & SRMMU_DIRTY; pte 221 arch/sparc/include/asm/pgtable_32.h static inline int pte_young(pte_t pte) pte 223 arch/sparc/include/asm/pgtable_32.h return pte_val(pte) & SRMMU_REF; pte 226 arch/sparc/include/asm/pgtable_32.h static inline int pte_special(pte_t pte) pte 231 arch/sparc/include/asm/pgtable_32.h static inline pte_t pte_wrprotect(pte_t pte) pte 233 arch/sparc/include/asm/pgtable_32.h return __pte(pte_val(pte) & ~SRMMU_WRITE); pte 236 arch/sparc/include/asm/pgtable_32.h static inline pte_t pte_mkclean(pte_t pte) pte 238 arch/sparc/include/asm/pgtable_32.h return __pte(pte_val(pte) & ~SRMMU_DIRTY); pte 241 arch/sparc/include/asm/pgtable_32.h static inline pte_t pte_mkold(pte_t pte) pte 243 arch/sparc/include/asm/pgtable_32.h return __pte(pte_val(pte) & ~SRMMU_REF); pte 246 arch/sparc/include/asm/pgtable_32.h static inline pte_t pte_mkwrite(pte_t pte) pte 248 arch/sparc/include/asm/pgtable_32.h return __pte(pte_val(pte) | SRMMU_WRITE); pte 251 arch/sparc/include/asm/pgtable_32.h static inline pte_t pte_mkdirty(pte_t pte) pte 253 arch/sparc/include/asm/pgtable_32.h return __pte(pte_val(pte) | SRMMU_DIRTY); pte 256 arch/sparc/include/asm/pgtable_32.h static inline pte_t pte_mkyoung(pte_t pte) pte 258 arch/sparc/include/asm/pgtable_32.h return __pte(pte_val(pte) | SRMMU_REF); pte 261 arch/sparc/include/asm/pgtable_32.h #define pte_mkspecial(pte) (pte) pte 265 arch/sparc/include/asm/pgtable_32.h static inline unsigned long pte_pfn(pte_t pte) pte 267 arch/sparc/include/asm/pgtable_32.h if (srmmu_device_memory(pte_val(pte))) { pte 275 arch/sparc/include/asm/pgtable_32.h return (pte_val(pte) & SRMMU_PTE_PMASK) >> (PAGE_SHIFT-4); pte 278 arch/sparc/include/asm/pgtable_32.h #define pte_page(pte) pfn_to_page(pte_pfn(pte)) pte 306 arch/sparc/include/asm/pgtable_32.h static pte_t pte_modify(pte_t pte, pgprot_t newprot) __attribute_const__; pte 307 arch/sparc/include/asm/pgtable_32.h static inline pte_t pte_modify(pte_t pte, pgprot_t newprot) pte 309 arch/sparc/include/asm/pgtable_32.h return __pte((pte_val(pte) & SRMMU_CHG_MASK) | pte 335 arch/sparc/include/asm/pgtable_32.h #define pte_unmap(pte) do{}while(0) pte 369 arch/sparc/include/asm/pgtable_32.h #define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val(pte) }) pte 251 arch/sparc/include/asm/pgtable_64.h pte_t pte = pfn_pte(page_nr, pgprot); pte 253 arch/sparc/include/asm/pgtable_64.h return __pmd(pte_val(pte)); pte 259 arch/sparc/include/asm/pgtable_64.h static inline unsigned long pte_pfn(pte_t pte) pte 272 arch/sparc/include/asm/pgtable_64.h : "r" (pte_val(pte)), pte 280 arch/sparc/include/asm/pgtable_64.h static inline pte_t pte_modify(pte_t pte, pgprot_t prot) pte 335 arch/sparc/include/asm/pgtable_64.h return __pte((pte_val(pte) & mask) | (pgprot_val(prot) & ~mask)); pte 341 arch/sparc/include/asm/pgtable_64.h pte_t pte = __pte(pmd_val(pmd)); pte 343 arch/sparc/include/asm/pgtable_64.h pte = pte_modify(pte, newprot); pte 345 arch/sparc/include/asm/pgtable_64.h return __pmd(pte_val(pte)); pte 401 arch/sparc/include/asm/pgtable_64.h static inline pte_t pte_mkhuge(pte_t pte) pte 403 arch/sparc/include/asm/pgtable_64.h return __pte(pte_val(pte) | __pte_default_huge_mask()); pte 406 arch/sparc/include/asm/pgtable_64.h static inline bool is_default_hugetlb_pte(pte_t pte) pte 410 arch/sparc/include/asm/pgtable_64.h return (pte_val(pte) & mask) == mask; pte 426 arch/sparc/include/asm/pgtable_64.h pte_t pte = __pte(pmd_val(pmd)); pte 428 arch/sparc/include/asm/pgtable_64.h pte = pte_mkhuge(pte); pte 429 arch/sparc/include/asm/pgtable_64.h pte_val(pte) |= _PAGE_PMD_HUGE; pte 431 arch/sparc/include/asm/pgtable_64.h return __pmd(pte_val(pte)); pte 435 arch/sparc/include/asm/pgtable_64.h static inline bool is_hugetlb_pte(pte_t pte) pte 441 arch/sparc/include/asm/pgtable_64.h static inline pte_t pte_mkdirty(pte_t pte) pte 443 arch/sparc/include/asm/pgtable_64.h unsigned long val = pte_val(pte), tmp; pte 465 arch/sparc/include/asm/pgtable_64.h static inline pte_t pte_mkclean(pte_t pte) pte 467 arch/sparc/include/asm/pgtable_64.h unsigned long val = pte_val(pte), tmp; pte 489 arch/sparc/include/asm/pgtable_64.h static inline pte_t pte_mkwrite(pte_t pte) pte 491 arch/sparc/include/asm/pgtable_64.h unsigned long val = pte_val(pte), mask; pte 507 arch/sparc/include/asm/pgtable_64.h static inline pte_t pte_wrprotect(pte_t pte) pte 509 arch/sparc/include/asm/pgtable_64.h unsigned long val = pte_val(pte), tmp; pte 531 arch/sparc/include/asm/pgtable_64.h static inline pte_t pte_mkold(pte_t pte) pte 548 arch/sparc/include/asm/pgtable_64.h return __pte(pte_val(pte) & ~mask); pte 551 arch/sparc/include/asm/pgtable_64.h static inline pte_t pte_mkyoung(pte_t pte) pte 568 arch/sparc/include/asm/pgtable_64.h return __pte(pte_val(pte) | mask); pte 571 arch/sparc/include/asm/pgtable_64.h static inline pte_t pte_mkspecial(pte_t pte) pte 573 arch/sparc/include/asm/pgtable_64.h pte_val(pte) |= _PAGE_SPECIAL; pte 574 arch/sparc/include/asm/pgtable_64.h return pte; pte 577 arch/sparc/include/asm/pgtable_64.h static inline pte_t pte_mkmcd(pte_t pte) pte 579 arch/sparc/include/asm/pgtable_64.h pte_val(pte) |= _PAGE_MCD_4V; pte 580 arch/sparc/include/asm/pgtable_64.h return pte; pte 583 arch/sparc/include/asm/pgtable_64.h static inline pte_t pte_mknotmcd(pte_t pte) pte 585 arch/sparc/include/asm/pgtable_64.h pte_val(pte) &= ~_PAGE_MCD_4V; pte 586 arch/sparc/include/asm/pgtable_64.h return pte; pte 589 arch/sparc/include/asm/pgtable_64.h static inline unsigned long pte_young(pte_t pte) pte 604 arch/sparc/include/asm/pgtable_64.h return (pte_val(pte) & mask); pte 607 arch/sparc/include/asm/pgtable_64.h static inline unsigned long pte_dirty(pte_t pte) pte 622 arch/sparc/include/asm/pgtable_64.h return (pte_val(pte) & mask); pte 625 arch/sparc/include/asm/pgtable_64.h static inline unsigned long pte_write(pte_t pte) pte 640 arch/sparc/include/asm/pgtable_64.h return (pte_val(pte) & mask); pte 643 arch/sparc/include/asm/pgtable_64.h static inline unsigned long pte_exec(pte_t pte) pte 656 arch/sparc/include/asm/pgtable_64.h return (pte_val(pte) & mask); pte 659 arch/sparc/include/asm/pgtable_64.h static inline unsigned long pte_present(pte_t pte) pte 661 arch/sparc/include/asm/pgtable_64.h unsigned long val = pte_val(pte); pte 681 arch/sparc/include/asm/pgtable_64.h static inline unsigned long pte_special(pte_t pte) pte 683 arch/sparc/include/asm/pgtable_64.h return pte_val(pte) & _PAGE_SPECIAL; pte 688 arch/sparc/include/asm/pgtable_64.h pte_t pte = __pte(pmd_val(pmd)); pte 690 arch/sparc/include/asm/pgtable_64.h return pte_val(pte) & _PAGE_PMD_HUGE; pte 695 arch/sparc/include/asm/pgtable_64.h pte_t pte = __pte(pmd_val(pmd)); pte 697 arch/sparc/include/asm/pgtable_64.h return pte_pfn(pte); pte 703 arch/sparc/include/asm/pgtable_64.h pte_t pte = __pte(pmd_val(pmd)); pte 705 arch/sparc/include/asm/pgtable_64.h return pte_write(pte); pte 713 arch/sparc/include/asm/pgtable_64.h pte_t pte = __pte(pmd_val(pmd)); pte 715 arch/sparc/include/asm/pgtable_64.h return pte_dirty(pte); pte 720 arch/sparc/include/asm/pgtable_64.h pte_t pte = __pte(pmd_val(pmd)); pte 722 arch/sparc/include/asm/pgtable_64.h return pte_young(pte); pte 727 arch/sparc/include/asm/pgtable_64.h pte_t pte = __pte(pmd_val(pmd)); pte 729 arch/sparc/include/asm/pgtable_64.h return pte_val(pte) & _PAGE_PMD_HUGE; pte 734 arch/sparc/include/asm/pgtable_64.h pte_t pte = __pte(pmd_val(pmd)); pte 736 arch/sparc/include/asm/pgtable_64.h pte = pte_mkold(pte); pte 738 arch/sparc/include/asm/pgtable_64.h return __pmd(pte_val(pte)); pte 743 arch/sparc/include/asm/pgtable_64.h pte_t pte = __pte(pmd_val(pmd)); pte 745 arch/sparc/include/asm/pgtable_64.h pte = pte_wrprotect(pte); pte 747 arch/sparc/include/asm/pgtable_64.h return __pmd(pte_val(pte)); pte 752 arch/sparc/include/asm/pgtable_64.h pte_t pte = __pte(pmd_val(pmd)); pte 754 arch/sparc/include/asm/pgtable_64.h pte = pte_mkdirty(pte); pte 756 arch/sparc/include/asm/pgtable_64.h return __pmd(pte_val(pte)); pte 761 arch/sparc/include/asm/pgtable_64.h pte_t pte = __pte(pmd_val(pmd)); pte 763 arch/sparc/include/asm/pgtable_64.h pte = pte_mkclean(pte); pte 765 arch/sparc/include/asm/pgtable_64.h return __pmd(pte_val(pte)); pte 770 arch/sparc/include/asm/pgtable_64.h pte_t pte = __pte(pmd_val(pmd)); pte 772 arch/sparc/include/asm/pgtable_64.h pte = pte_mkyoung(pte); pte 774 arch/sparc/include/asm/pgtable_64.h return __pmd(pte_val(pte)); pte 779 arch/sparc/include/asm/pgtable_64.h pte_t pte = __pte(pmd_val(pmd)); pte 781 arch/sparc/include/asm/pgtable_64.h pte = pte_mkwrite(pte); pte 783 arch/sparc/include/asm/pgtable_64.h return __pmd(pte_val(pte)); pte 839 arch/sparc/include/asm/pgtable_64.h pte_t pte = __pte(pmd_val(pmd)); pte 842 arch/sparc/include/asm/pgtable_64.h pfn = pte_pfn(pte); pte 849 arch/sparc/include/asm/pgtable_64.h pte_t pte = __pte(pud_val(pud)); pte 852 arch/sparc/include/asm/pgtable_64.h pfn = pte_pfn(pte); pte 872 arch/sparc/include/asm/pgtable_64.h pte_t pte = __pte(pud_val(pud)); pte 874 arch/sparc/include/asm/pgtable_64.h return pte_val(pte) & _PAGE_PMD_HUGE; pte 879 arch/sparc/include/asm/pgtable_64.h pte_t pte = __pte(pud_val(pud)); pte 881 arch/sparc/include/asm/pgtable_64.h return pte_pfn(pte); pte 885 arch/sparc/include/asm/pgtable_64.h #define pte_none(pte) (!pte_val(pte)) pte 913 arch/sparc/include/asm/pgtable_64.h #define pte_unmap(pte) do { } while (0) pte 948 arch/sparc/include/asm/pgtable_64.h pte_t *ptep, pte_t pte, int fullmm) pte 952 arch/sparc/include/asm/pgtable_64.h *ptep = pte; pte 956 arch/sparc/include/asm/pgtable_64.h #define set_pte_at(mm,addr,ptep,pte) \ pte 957 arch/sparc/include/asm/pgtable_64.h __set_pte_at((mm), (addr), (ptep), (pte), 0) pte 968 arch/sparc/include/asm/pgtable_64.h #define move_pte(pte, prot, old_addr, new_addr) \ pte 970 arch/sparc/include/asm/pgtable_64.h pte_t newpte = (pte); \ pte 971 arch/sparc/include/asm/pgtable_64.h if (tlb_type != hypervisor && pte_present(pte)) { \ pte 972 arch/sparc/include/asm/pgtable_64.h unsigned long this_pfn = pte_pfn(pte); \ pte 1018 arch/sparc/include/asm/pgtable_64.h #define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val(pte) }) pte 1035 arch/sparc/include/asm/pgtable_64.h unsigned long addr, pte_t pte); pte 1044 arch/sparc/include/asm/pgtable_64.h pte_t pte, pte_t oldpte) pte 1053 arch/sparc/include/asm/pgtable_64.h if (adi_state.enabled && (pte_val(pte) & _PAGE_MCD_4V)) pte 1054 arch/sparc/include/asm/pgtable_64.h adi_restore_tags(mm, vma, addr, pte); pte 1104 arch/sparc/include/asm/pgtable_64.h static inline bool pte_access_permitted(pte_t pte, bool write) pte 1118 arch/sparc/include/asm/pgtable_64.h return (pte_val(pte) & (prot | _PAGE_SPECIAL)) == prot; pte 328 arch/sparc/kernel/adi_64.c unsigned long addr, pte_t pte) pte 343 arch/sparc/kernel/adi_64.c paddr = pte_val(pte) & _PAGE_PADDR_4V; pte 300 arch/sparc/kernel/signal32.c pte_t *ptep, pte; pte 329 arch/sparc/kernel/signal32.c pte = *ptep; pte 330 arch/sparc/kernel/signal32.c if (!pte_present(pte)) pte 333 arch/sparc/kernel/signal32.c paddr = (unsigned long) page_address(pte_page(pte)); pte 85 arch/sparc/mm/fault_64.c pte_t *ptep, pte; pte 115 arch/sparc/mm/fault_64.c pte = *ptep; pte 116 arch/sparc/mm/fault_64.c if (pte_present(pte)) { pte 117 arch/sparc/mm/fault_64.c pa = (pte_pfn(pte) << PAGE_SHIFT); pte 185 arch/sparc/mm/hugetlbpage.c pte_t pte; pte 187 arch/sparc/mm/hugetlbpage.c pte = hugepage_shift_to_tte(entry, shift); pte 193 arch/sparc/mm/hugetlbpage.c return pte_mkmcd(pte); pte 195 arch/sparc/mm/hugetlbpage.c return pte_mknotmcd(pte); pte 197 arch/sparc/mm/hugetlbpage.c return pte; pte 267 arch/sparc/mm/hugetlbpage.c static unsigned long huge_tte_to_size(pte_t pte) pte 269 arch/sparc/mm/hugetlbpage.c unsigned long size = 1UL << huge_tte_to_shift(pte); pte 269 arch/sparc/mm/init_64.c static inline void tsb_insert(struct tsb *ent, unsigned long tag, unsigned long pte) pte 276 arch/sparc/mm/init_64.c __tsb_insert(tsb_addr, tag, pte); pte 421 arch/sparc/mm/init_64.c pte_t pte = *ptep; pte 424 arch/sparc/mm/init_64.c unsigned long pfn = pte_pfn(pte); pte 433 arch/sparc/mm/init_64.c if (!pte_accessible(mm, pte)) pte 452 arch/sparc/mm/init_64.c pte_val(pte) &= ~mask; pte 453 arch/sparc/mm/init_64.c pte_val(pte) |= (address & mask); pte 458 arch/sparc/mm/init_64.c pte_val(pte) |= (address & (1UL << REAL_HPAGE_SHIFT)); pte 463 arch/sparc/mm/init_64.c REAL_HPAGE_SHIFT, address, pte_val(pte)); pte 470 arch/sparc/mm/init_64.c address, pte_val(pte)); pte 679 arch/sparc/mm/init_64.c unsigned long pte, pte 682 arch/sparc/mm/init_64.c unsigned long ret = sun4v_mmu_map_perm_addr(vaddr, 0, pte, mmu); pte 686 arch/sparc/mm/init_64.c "errors with %lx\n", vaddr, 0, pte, mmu, ret); pte 1658 arch/sparc/mm/init_64.c pte_t *pte; pte 1691 arch/sparc/mm/init_64.c pte = pte_offset_kernel(pmd, addr); pte 1692 arch/sparc/mm/init_64.c if (pte_none(*pte)) pte 1695 arch/sparc/mm/init_64.c return pfn_valid(pte_pfn(*pte)); pte 1805 arch/sparc/mm/init_64.c pte_t *pte; pte 1849 arch/sparc/mm/init_64.c pte = pte_offset_kernel(pmd, vstart); pte 1855 arch/sparc/mm/init_64.c pte_val(*pte) = (paddr | pgprot_val(prot)); pte 1859 arch/sparc/mm/init_64.c pte++; pte 2614 arch/sparc/mm/init_64.c unsigned long pte; pte 2626 arch/sparc/mm/init_64.c pte = pmd_val(*pmd); pte 2627 arch/sparc/mm/init_64.c if (!(pte & _PAGE_VALID)) { pte 2807 arch/sparc/mm/init_64.c pte_t pte; pte 2809 arch/sparc/mm/init_64.c pte_val(pte) = page | pgprot_val(pgprot_noncached(prot)); pte 2810 arch/sparc/mm/init_64.c pte_val(pte) |= (((unsigned long)space) << 32); pte 2811 arch/sparc/mm/init_64.c pte_val(pte) |= pte_sz_bits(page_size); pte 2813 arch/sparc/mm/init_64.c return pte; pte 2893 arch/sparc/mm/init_64.c pte_t *pte = NULL; pte 2896 arch/sparc/mm/init_64.c pte = (pte_t *) page_address(page); pte 2898 arch/sparc/mm/init_64.c return pte; pte 2913 arch/sparc/mm/init_64.c void pte_free_kernel(struct mm_struct *mm, pte_t *pte) pte 2915 arch/sparc/mm/init_64.c free_page((unsigned long)pte); pte 2918 arch/sparc/mm/init_64.c static void __pte_free(pgtable_t pte) pte 2920 arch/sparc/mm/init_64.c struct page *page = virt_to_page(pte); pte 2926 arch/sparc/mm/init_64.c void pte_free(struct mm_struct *mm, pgtable_t pte) pte 2928 arch/sparc/mm/init_64.c __pte_free(pte); pte 2943 arch/sparc/mm/init_64.c unsigned long pte, flags; pte 2950 arch/sparc/mm/init_64.c pte = pmd_val(entry); pte 2953 arch/sparc/mm/init_64.c if (!(pte & _PAGE_VALID)) pte 2957 arch/sparc/mm/init_64.c pte |= (addr & (1UL << REAL_HPAGE_SHIFT)); pte 2965 arch/sparc/mm/init_64.c addr, pte); pte 42 arch/sparc/mm/leon_mm.c unsigned int lvl, pte, paddrbase; pte 75 arch/sparc/mm/leon_mm.c pte = pgd; pte 98 arch/sparc/mm/leon_mm.c pte = pmd; pte 126 arch/sparc/mm/leon_mm.c pte = ped; pte 149 arch/sparc/mm/leon_mm.c pte = ptr; pte 161 arch/sparc/mm/leon_mm.c (vaddr & ~(-1 << LEON_PTE_SH)) | ((pte & ~0xff) << 4); pte 165 arch/sparc/mm/leon_mm.c (vaddr & ~(-1 << LEON_PMD_SH)) | ((pte & ~0xff) << 4); pte 169 arch/sparc/mm/leon_mm.c (vaddr & ~(-1 << LEON_PGD_SH)) | ((pte & ~0xff) << 4); pte 180 arch/sparc/mm/leon_mm.c return pte; pte 112 arch/sparc/mm/srmmu.c pte_t pte; pte 114 arch/sparc/mm/srmmu.c pte = __pte((SRMMU_ET_PTD | (__nocache_pa(pgdp) >> 4))); pte 115 arch/sparc/mm/srmmu.c set_pte((pte_t *)ctxp, pte); pte 164 arch/sparc/mm/srmmu.c void *pte; pte 166 arch/sparc/mm/srmmu.c pte = __nocache_va((dir->pmdv[0] & SRMMU_PTD_PMASK) << 4); pte 167 arch/sparc/mm/srmmu.c return (pte_t *) pte + pte 300 arch/sparc/mm/srmmu.c pte_t *pte; pte 333 arch/sparc/mm/srmmu.c pte = pte_offset_kernel(__nocache_fix(pmd), vaddr); pte 340 arch/sparc/mm/srmmu.c set_pte(__nocache_fix(pte), __pte(pteval)); pte 375 arch/sparc/mm/srmmu.c unsigned long pte; pte 378 arch/sparc/mm/srmmu.c if ((pte = (unsigned long)pte_alloc_one_kernel(mm)) == 0) pte 380 arch/sparc/mm/srmmu.c page = pfn_to_page(__nocache_pa(pte) >> PAGE_SHIFT); pte 388 arch/sparc/mm/srmmu.c void pte_free(struct mm_struct *mm, pgtable_t pte) pte 392 arch/sparc/mm/srmmu.c pgtable_pte_page_dtor(pte); pte 393 arch/sparc/mm/srmmu.c p = (unsigned long)page_address(pte); /* Cached address (for test) */ pte 396 arch/sparc/mm/srmmu.c p = page_to_pfn(pte) << PAGE_SHIFT; /* Physical address */ pte 910 arch/sparc/mm/srmmu.c pte_t *pte; pte 971 arch/sparc/mm/srmmu.c pte = pte_offset_kernel(pmd, PKMAP_BASE); pte 972 arch/sparc/mm/srmmu.c pkmap_page_table = pte; pte 150 arch/sparc/mm/tlb.c pte_t *pte; pte 152 arch/sparc/mm/tlb.c pte = pte_offset_map(&pmd, vaddr); pte 155 arch/sparc/mm/tlb.c if (pte_val(*pte) & _PAGE_VALID) { pte 156 arch/sparc/mm/tlb.c bool exec = pte_exec(*pte); pte 160 arch/sparc/mm/tlb.c pte++; pte 163 arch/sparc/mm/tlb.c pte_unmap(pte); pte 37 arch/um/include/asm/page.h typedef struct { unsigned long pte; } pte_t; pte 40 arch/um/include/asm/page.h #define pte_val(p) ((p).pte) pte 42 arch/um/include/asm/page.h #define pte_get_bits(p, bits) ((p).pte & (bits)) pte 43 arch/um/include/asm/page.h #define pte_set_bits(p, bits) ((p).pte |= (bits)) pte 44 arch/um/include/asm/page.h #define pte_clear_bits(p, bits) ((p).pte &= ~(bits)) pte 45 arch/um/include/asm/page.h #define pte_copy(to, from) ({ (to).pte = (from).pte; }) pte 46 arch/um/include/asm/page.h #define pte_is_zero(p) (!((p).pte & ~_PAGE_NEWPAGE)) pte 48 arch/um/include/asm/page.h ({ (p).pte = (phys) | pgprot_val(prot); }) pte 57 arch/um/include/asm/page.h typedef struct { unsigned long pte; } pte_t; pte 66 arch/um/include/asm/page.h #define pte_val(x) ((x).pte) pte 69 arch/um/include/asm/page.h #define pte_get_bits(p, bits) ((p).pte & (bits)) pte 70 arch/um/include/asm/page.h #define pte_set_bits(p, bits) ((p).pte |= (bits)) pte 71 arch/um/include/asm/page.h #define pte_clear_bits(p, bits) ((p).pte &= ~(bits)) pte 72 arch/um/include/asm/page.h #define pte_copy(to, from) ((to).pte = (from).pte) pte 73 arch/um/include/asm/page.h #define pte_is_zero(p) (!((p).pte & ~_PAGE_NEWPAGE)) pte 74 arch/um/include/asm/page.h #define pte_set_val(p, phys, prot) (p).pte = (phys | pgprot_val(prot)) pte 15 arch/um/include/asm/pgalloc.h #define pmd_populate_kernel(mm, pmd, pte) \ pte 16 arch/um/include/asm/pgalloc.h set_pmd(pmd, __pmd(_PAGE_TABLE + (unsigned long) __pa(pte))) pte 18 arch/um/include/asm/pgalloc.h #define pmd_populate(mm, pmd, pte) \ pte 20 arch/um/include/asm/pgalloc.h ((unsigned long long)page_to_pfn(pte) << \ pte 30 arch/um/include/asm/pgalloc.h #define __pte_free_tlb(tlb,pte, address) \ pte 32 arch/um/include/asm/pgalloc.h pgtable_pte_page_dtor(pte); \ pte 33 arch/um/include/asm/pgalloc.h tlb_remove_page((tlb),(pte)); \ pte 97 arch/um/include/asm/pgtable-3level.h static inline unsigned long pte_pfn(pte_t pte) pte 99 arch/um/include/asm/pgtable-3level.h return phys_to_pfn(pte_val(pte)); pte 104 arch/um/include/asm/pgtable-3level.h pte_t pte; pte 107 arch/um/include/asm/pgtable-3level.h pte_set_val(pte, phys, pgprot); pte 108 arch/um/include/asm/pgtable-3level.h return pte; pte 121 arch/um/include/asm/pgtable.h static inline int pte_none(pte_t pte) pte 123 arch/um/include/asm/pgtable.h return pte_is_zero(pte); pte 130 arch/um/include/asm/pgtable.h static inline int pte_read(pte_t pte) pte 132 arch/um/include/asm/pgtable.h return((pte_get_bits(pte, _PAGE_USER)) && pte 133 arch/um/include/asm/pgtable.h !(pte_get_bits(pte, _PAGE_PROTNONE))); pte 136 arch/um/include/asm/pgtable.h static inline int pte_exec(pte_t pte){ pte 137 arch/um/include/asm/pgtable.h return((pte_get_bits(pte, _PAGE_USER)) && pte 138 arch/um/include/asm/pgtable.h !(pte_get_bits(pte, _PAGE_PROTNONE))); pte 141 arch/um/include/asm/pgtable.h static inline int pte_write(pte_t pte) pte 143 arch/um/include/asm/pgtable.h return((pte_get_bits(pte, _PAGE_RW)) && pte 144 arch/um/include/asm/pgtable.h !(pte_get_bits(pte, _PAGE_PROTNONE))); pte 147 arch/um/include/asm/pgtable.h static inline int pte_dirty(pte_t pte) pte 149 arch/um/include/asm/pgtable.h return pte_get_bits(pte, _PAGE_DIRTY); pte 152 arch/um/include/asm/pgtable.h static inline int pte_young(pte_t pte) pte 154 arch/um/include/asm/pgtable.h return pte_get_bits(pte, _PAGE_ACCESSED); pte 157 arch/um/include/asm/pgtable.h static inline int pte_newpage(pte_t pte) pte 159 arch/um/include/asm/pgtable.h return pte_get_bits(pte, _PAGE_NEWPAGE); pte 162 arch/um/include/asm/pgtable.h static inline int pte_newprot(pte_t pte) pte 164 arch/um/include/asm/pgtable.h return(pte_present(pte) && (pte_get_bits(pte, _PAGE_NEWPROT))); pte 167 arch/um/include/asm/pgtable.h static inline int pte_special(pte_t pte) pte 178 arch/um/include/asm/pgtable.h static inline pte_t pte_mknewprot(pte_t pte) pte 180 arch/um/include/asm/pgtable.h pte_set_bits(pte, _PAGE_NEWPROT); pte 181 arch/um/include/asm/pgtable.h return(pte); pte 184 arch/um/include/asm/pgtable.h static inline pte_t pte_mkclean(pte_t pte) pte 186 arch/um/include/asm/pgtable.h pte_clear_bits(pte, _PAGE_DIRTY); pte 187 arch/um/include/asm/pgtable.h return(pte); pte 190 arch/um/include/asm/pgtable.h static inline pte_t pte_mkold(pte_t pte) pte 192 arch/um/include/asm/pgtable.h pte_clear_bits(pte, _PAGE_ACCESSED); pte 193 arch/um/include/asm/pgtable.h return(pte); pte 196 arch/um/include/asm/pgtable.h static inline pte_t pte_wrprotect(pte_t pte) pte 198 arch/um/include/asm/pgtable.h if (likely(pte_get_bits(pte, _PAGE_RW))) pte 199 arch/um/include/asm/pgtable.h pte_clear_bits(pte, _PAGE_RW); pte 201 arch/um/include/asm/pgtable.h return pte; pte 202 arch/um/include/asm/pgtable.h return(pte_mknewprot(pte)); pte 205 arch/um/include/asm/pgtable.h static inline pte_t pte_mkread(pte_t pte) pte 207 arch/um/include/asm/pgtable.h if (unlikely(pte_get_bits(pte, _PAGE_USER))) pte 208 arch/um/include/asm/pgtable.h return pte; pte 209 arch/um/include/asm/pgtable.h pte_set_bits(pte, _PAGE_USER); pte 210 arch/um/include/asm/pgtable.h return(pte_mknewprot(pte)); pte 213 arch/um/include/asm/pgtable.h static inline pte_t pte_mkdirty(pte_t pte) pte 215 arch/um/include/asm/pgtable.h pte_set_bits(pte, _PAGE_DIRTY); pte 216 arch/um/include/asm/pgtable.h return(pte); pte 219 arch/um/include/asm/pgtable.h static inline pte_t pte_mkyoung(pte_t pte) pte 221 arch/um/include/asm/pgtable.h pte_set_bits(pte, _PAGE_ACCESSED); pte 222 arch/um/include/asm/pgtable.h return(pte); pte 225 arch/um/include/asm/pgtable.h static inline pte_t pte_mkwrite(pte_t pte) pte 227 arch/um/include/asm/pgtable.h if (unlikely(pte_get_bits(pte, _PAGE_RW))) pte 228 arch/um/include/asm/pgtable.h return pte; pte 229 arch/um/include/asm/pgtable.h pte_set_bits(pte, _PAGE_RW); pte 230 arch/um/include/asm/pgtable.h return(pte_mknewprot(pte)); pte 233 arch/um/include/asm/pgtable.h static inline pte_t pte_mkuptodate(pte_t pte) pte 235 arch/um/include/asm/pgtable.h pte_clear_bits(pte, _PAGE_NEWPAGE); pte 236 arch/um/include/asm/pgtable.h if(pte_present(pte)) pte 237 arch/um/include/asm/pgtable.h pte_clear_bits(pte, _PAGE_NEWPROT); pte 238 arch/um/include/asm/pgtable.h return(pte); pte 241 arch/um/include/asm/pgtable.h static inline pte_t pte_mknewpage(pte_t pte) pte 243 arch/um/include/asm/pgtable.h pte_set_bits(pte, _PAGE_NEWPAGE); pte 244 arch/um/include/asm/pgtable.h return(pte); pte 247 arch/um/include/asm/pgtable.h static inline pte_t pte_mkspecial(pte_t pte) pte 249 arch/um/include/asm/pgtable.h return(pte); pte 288 arch/um/include/asm/pgtable.h ({ pte_t pte; \ pte 290 arch/um/include/asm/pgtable.h pte_set_val(pte, page_to_phys(page), (pgprot)); \ pte 291 arch/um/include/asm/pgtable.h if (pte_present(pte)) \ pte 292 arch/um/include/asm/pgtable.h pte_mknewprot(pte_mknewpage(pte)); \ pte 293 arch/um/include/asm/pgtable.h pte;}) pte 295 arch/um/include/asm/pgtable.h static inline pte_t pte_modify(pte_t pte, pgprot_t newprot) pte 297 arch/um/include/asm/pgtable.h pte_set_val(pte, (pte_val(pte) & _PAGE_CHG_MASK), newprot); pte 298 arch/um/include/asm/pgtable.h return pte; pte 344 arch/um/include/asm/pgtable.h #define pte_unmap(pte) do { } while (0) pte 357 arch/um/include/asm/pgtable.h #define __pte_to_swp_entry(pte) \ pte 358 arch/um/include/asm/pgtable.h ((swp_entry_t) { pte_val(pte_mkuptodate(pte)) }) pte 68 arch/um/kernel/mem.c pte_t *pte = (pte_t *) memblock_alloc_low(PAGE_SIZE, pte 70 arch/um/kernel/mem.c if (!pte) pte 75 arch/um/kernel/mem.c (unsigned long) __pa(pte))); pte 76 arch/um/kernel/mem.c if (pte != pte_offset_kernel(pmd, 0)) pte 129 arch/um/kernel/mem.c pte_t *pte; pte 149 arch/um/kernel/mem.c pte = pte_offset_kernel(pmd, vaddr); pte 150 arch/um/kernel/mem.c pte_set_val(*pte, p, PAGE_READONLY); pte 24 arch/um/kernel/skas/mmu.c pte_t *pte; pte 35 arch/um/kernel/skas/mmu.c pte = pte_alloc_map(mm, pmd, proc); pte 36 arch/um/kernel/skas/mmu.c if (!pte) pte 39 arch/um/kernel/skas/mmu.c *pte = mk_pte(virt_to_page(kernel), __pgprot(_PAGE_PRESENT)); pte 40 arch/um/kernel/skas/mmu.c *pte = pte_mkread(*pte); pte 127 arch/um/kernel/skas/mmu.c pte_t *pte; pte 129 arch/um/kernel/skas/mmu.c pte = virt_to_pte(mm, STUB_CODE); pte 130 arch/um/kernel/skas/mmu.c if (pte != NULL) pte 131 arch/um/kernel/skas/mmu.c pte_clear(mm, STUB_CODE, pte); pte 133 arch/um/kernel/skas/mmu.c pte = virt_to_pte(mm, STUB_DATA); pte 134 arch/um/kernel/skas/mmu.c if (pte == NULL) pte 137 arch/um/kernel/skas/mmu.c pte_clear(mm, STUB_DATA, pte); pte 43 arch/um/kernel/skas/uaccess.c pte_t *pte = virt_to_pte(current->mm, virt); pte 46 arch/um/kernel/skas/uaccess.c if ((pte == NULL) || !pte_present(*pte) || pte 47 arch/um/kernel/skas/uaccess.c (is_write && !pte_write(*pte))) { pte 51 arch/um/kernel/skas/uaccess.c pte = virt_to_pte(current->mm, virt); pte 53 arch/um/kernel/skas/uaccess.c if (!pte_present(*pte)) pte 54 arch/um/kernel/skas/uaccess.c pte = NULL; pte 56 arch/um/kernel/skas/uaccess.c return pte; pte 63 arch/um/kernel/skas/uaccess.c pte_t *pte; pte 66 arch/um/kernel/skas/uaccess.c pte = maybe_map(addr, is_write); pte 67 arch/um/kernel/skas/uaccess.c if (pte == NULL) pte 70 arch/um/kernel/skas/uaccess.c page = pte_page(*pte); pte 225 arch/um/kernel/tlb.c pte_t *pte; pte 228 arch/um/kernel/tlb.c pte = pte_offset_kernel(pmd, addr); pte 233 arch/um/kernel/tlb.c r = pte_read(*pte); pte 234 arch/um/kernel/tlb.c w = pte_write(*pte); pte 235 arch/um/kernel/tlb.c x = pte_exec(*pte); pte 236 arch/um/kernel/tlb.c if (!pte_young(*pte)) { pte 239 arch/um/kernel/tlb.c } else if (!pte_dirty(*pte)) pte 244 arch/um/kernel/tlb.c if (hvc->force || pte_newpage(*pte)) { pte 245 arch/um/kernel/tlb.c if (pte_present(*pte)) { pte 246 arch/um/kernel/tlb.c if (pte_newpage(*pte)) pte 247 arch/um/kernel/tlb.c ret = add_mmap(addr, pte_val(*pte) & PAGE_MASK, pte 251 arch/um/kernel/tlb.c } else if (pte_newprot(*pte)) pte 253 arch/um/kernel/tlb.c *pte = pte_mkuptodate(*pte); pte 254 arch/um/kernel/tlb.c } while (pte++, addr += PAGE_SIZE, ((addr < end) && !ret)); pte 343 arch/um/kernel/tlb.c pte_t *pte; pte 399 arch/um/kernel/tlb.c pte = pte_offset_kernel(pmd, addr); pte 400 arch/um/kernel/tlb.c if (!pte_present(*pte) || pte_newpage(*pte)) { pte 406 arch/um/kernel/tlb.c if (pte_present(*pte)) pte 407 arch/um/kernel/tlb.c err = add_mmap(addr, pte_val(*pte) & PAGE_MASK, pte 410 arch/um/kernel/tlb.c else if (pte_newprot(*pte)) { pte 429 arch/um/kernel/tlb.c pte_t *pte; pte 448 arch/um/kernel/tlb.c pte = pte_offset_kernel(pmd, address); pte 450 arch/um/kernel/tlb.c r = pte_read(*pte); pte 451 arch/um/kernel/tlb.c w = pte_write(*pte); pte 452 arch/um/kernel/tlb.c x = pte_exec(*pte); pte 453 arch/um/kernel/tlb.c if (!pte_young(*pte)) { pte 456 arch/um/kernel/tlb.c } else if (!pte_dirty(*pte)) { pte 463 arch/um/kernel/tlb.c if (pte_newpage(*pte)) { pte 464 arch/um/kernel/tlb.c if (pte_present(*pte)) { pte 468 arch/um/kernel/tlb.c fd = phys_mapping(pte_val(*pte) & PAGE_MASK, &offset); pte 474 arch/um/kernel/tlb.c else if (pte_newprot(*pte)) pte 484 arch/um/kernel/tlb.c *pte = pte_mkuptodate(*pte); pte 33 arch/um/kernel/trap.c pte_t *pte; pte 109 arch/um/kernel/trap.c pte = pte_offset_kernel(pmd, address); pte 110 arch/um/kernel/trap.c } while (!pte_present(*pte)); pte 121 arch/um/kernel/trap.c WARN_ON(!pte_young(*pte) || (is_write && !pte_dirty(*pte))); pte 36 arch/unicore32/include/asm/cpu-single.h extern void cpu_set_pte(pte_t *ptep, pte_t pte); pte 34 arch/unicore32/include/asm/page.h typedef struct { unsigned long pte; } pte_t; pte 38 arch/unicore32/include/asm/page.h #define pte_val(x) ((x).pte) pte 36 arch/unicore32/include/asm/pgalloc.h pte_t *pte = __pte_alloc_one_kernel(mm); pte 38 arch/unicore32/include/asm/pgalloc.h if (pte) pte 39 arch/unicore32/include/asm/pgalloc.h clean_dcache_area(pte, PTRS_PER_PTE * sizeof(pte_t)); pte 41 arch/unicore32/include/asm/pgalloc.h return pte; pte 47 arch/unicore32/include/asm/pgalloc.h struct page *pte; pte 49 arch/unicore32/include/asm/pgalloc.h pte = __pte_alloc_one(mm, GFP_PGTABLE_USER); pte 50 arch/unicore32/include/asm/pgalloc.h if (!pte) pte 52 arch/unicore32/include/asm/pgalloc.h if (!PageHighMem(pte)) pte 53 arch/unicore32/include/asm/pgalloc.h clean_pte_table(page_address(pte)); pte 54 arch/unicore32/include/asm/pgalloc.h return pte; pte 50 arch/unicore32/include/asm/pgtable.h #define pte_ERROR(pte) __pte_error(__FILE__, __LINE__, pte_val(pte)) pte 150 arch/unicore32/include/asm/pgtable.h #define pte_pfn(pte) (pte_val(pte) >> PAGE_SHIFT) pte 154 arch/unicore32/include/asm/pgtable.h #define pte_none(pte) (!pte_val(pte)) pte 156 arch/unicore32/include/asm/pgtable.h #define pte_page(pte) (pfn_to_page(pte_pfn(pte))) pte 162 arch/unicore32/include/asm/pgtable.h #define pte_unmap(pte) do { } while (0) pte 164 arch/unicore32/include/asm/pgtable.h #define set_pte(ptep, pte) cpu_set_pte(ptep, pte) pte 175 arch/unicore32/include/asm/pgtable.h #define pte_present(pte) (pte_val(pte) & PTE_PRESENT) pte 176 arch/unicore32/include/asm/pgtable.h #define pte_write(pte) (pte_val(pte) & PTE_WRITE) pte 177 arch/unicore32/include/asm/pgtable.h #define pte_dirty(pte) (pte_val(pte) & PTE_DIRTY) pte 178 arch/unicore32/include/asm/pgtable.h #define pte_young(pte) (pte_val(pte) & PTE_YOUNG) pte 179 arch/unicore32/include/asm/pgtable.h #define pte_exec(pte) (pte_val(pte) & PTE_EXEC) pte 180 arch/unicore32/include/asm/pgtable.h #define pte_special(pte) (0) pte 183 arch/unicore32/include/asm/pgtable.h static inline pte_t pte_##fn(pte_t pte) { pte_val(pte) op; return pte; } pte 192 arch/unicore32/include/asm/pgtable.h static inline pte_t pte_mkspecial(pte_t pte) { return pte; } pte 239 arch/unicore32/include/asm/pgtable.h static inline pte_t pte_modify(pte_t pte, pgprot_t newprot) pte 242 arch/unicore32/include/asm/pgtable.h pte_val(pte) = (pte_val(pte) & ~mask) | (pgprot_val(newprot) & mask); pte 243 arch/unicore32/include/asm/pgtable.h return pte; pte 271 arch/unicore32/include/asm/pgtable.h #define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val(pte) }) pte 16 arch/unicore32/include/asm/tlb.h #define __pte_free_tlb(tlb, pte, addr) \ pte 18 arch/unicore32/include/asm/tlb.h pgtable_pte_page_dtor(pte); \ pte 19 arch/unicore32/include/asm/tlb.h tlb_remove_page((tlb), (pte)); \ pte 75 arch/unicore32/kernel/hibernate.c pte_t *pte; pte 99 arch/unicore32/kernel/hibernate.c pte = resume_one_page_table_init(pmd); pte 100 arch/unicore32/kernel/hibernate.c if (!pte) pte 103 arch/unicore32/kernel/hibernate.c max_pte = pte + PTRS_PER_PTE; pte 104 arch/unicore32/kernel/hibernate.c for (; pte < max_pte; pte++, pfn++) { pte 108 arch/unicore32/kernel/hibernate.c set_pte(pte, pfn_pte(pfn, PAGE_KERNEL_EXEC)); pte 51 arch/unicore32/mm/fault.c pte_t *pte; pte 77 arch/unicore32/mm/fault.c pte = pte_offset_map(pmd, addr); pte 78 arch/unicore32/mm/fault.c printk(", *pte=%08lx", pte_val(*pte)); pte 79 arch/unicore32/mm/fault.c pte_unmap(pte); pte 146 arch/unicore32/mm/mmu.c pte_t *pte = memblock_alloc(size, size); pte 148 arch/unicore32/mm/mmu.c if (!pte) pte 152 arch/unicore32/mm/mmu.c __pmd_populate(pmd, __pa(pte) | prot); pte 162 arch/unicore32/mm/mmu.c pte_t *pte = early_pte_alloc(pmd, addr, type->prot_l1); pte 164 arch/unicore32/mm/mmu.c set_pte(pte, pfn_pte(pfn, __pgprot(type->prot_pte))); pte 166 arch/unicore32/mm/mmu.c } while (pte++, addr += PAGE_SIZE, addr != end); pte 79 arch/unicore32/mm/pgd.c pgtable_t pte; pte 94 arch/unicore32/mm/pgd.c pte = pmd_pgtable(*pmd); pte 96 arch/unicore32/mm/pgd.c pte_free(mm, pte); pte 158 arch/x86/include/asm/fixmap.h void __native_set_fixmap(enum fixed_addresses idx, pte_t pte); pte 156 arch/x86/include/asm/kexec.h pte_t *pte; pte 395 arch/x86/include/asm/kvm_host.h u64 *spte, const void *pte); pte 1559 arch/x86/include/asm/kvm_host.h int kvm_set_spte_hva(struct kvm *kvm, unsigned long hva, pte_t pte); pte 371 arch/x86/include/asm/paravirt.h return (pte_t) { .pte = ret }; pte 374 arch/x86/include/asm/paravirt.h static inline pteval_t pte_val(pte_t pte) pte 380 arch/x86/include/asm/paravirt.h pte.pte, (u64)pte.pte >> 32); pte 382 arch/x86/include/asm/paravirt.h ret = PVOP_CALLEE1(pteval_t, mmu.pte_val, pte.pte); pte 420 arch/x86/include/asm/paravirt.h return (pte_t) { .pte = ret }; pte 424 arch/x86/include/asm/paravirt.h pte_t *ptep, pte_t old_pte, pte_t pte) pte 429 arch/x86/include/asm/paravirt.h pv_ops.mmu.ptep_modify_prot_commit(vma, addr, ptep, pte); pte 432 arch/x86/include/asm/paravirt.h vma, addr, ptep, pte.pte); pte 435 arch/x86/include/asm/paravirt.h static inline void set_pte(pte_t *ptep, pte_t pte) pte 438 arch/x86/include/asm/paravirt.h PVOP_VCALL3(mmu.set_pte, ptep, pte.pte, (u64)pte.pte >> 32); pte 440 arch/x86/include/asm/paravirt.h PVOP_VCALL2(mmu.set_pte, ptep, pte.pte); pte 444 arch/x86/include/asm/paravirt.h pte_t *ptep, pte_t pte) pte 448 arch/x86/include/asm/paravirt.h pv_ops.mmu.set_pte_at(mm, addr, ptep, pte); pte 450 arch/x86/include/asm/paravirt.h PVOP_VCALL4(mmu.set_pte_at, mm, addr, ptep, pte.pte); pte 570 arch/x86/include/asm/paravirt.h static inline void set_pte_atomic(pte_t *ptep, pte_t pte) pte 572 arch/x86/include/asm/paravirt.h PVOP_VCALL3(mmu.set_pte_atomic, ptep, pte.pte, pte.pte >> 32); pte 586 arch/x86/include/asm/paravirt.h static inline void set_pte_atomic(pte_t *ptep, pte_t pte) pte 588 arch/x86/include/asm/paravirt.h set_pte(ptep, pte); pte 256 arch/x86/include/asm/paravirt_types.h pte_t *ptep, pte_t pte); pte 55 arch/x86/include/asm/pgalloc.h extern void ___pte_free_tlb(struct mmu_gather *tlb, struct page *pte); pte 57 arch/x86/include/asm/pgalloc.h static inline void __pte_free_tlb(struct mmu_gather *tlb, struct page *pte, pte 60 arch/x86/include/asm/pgalloc.h ___pte_free_tlb(tlb, pte); pte 64 arch/x86/include/asm/pgalloc.h pmd_t *pmd, pte_t *pte) pte 66 arch/x86/include/asm/pgalloc.h paravirt_alloc_pte(mm, __pa(pte) >> PAGE_SHIFT); pte 67 arch/x86/include/asm/pgalloc.h set_pmd(pmd, __pmd(__pa(pte) | _PAGE_TABLE)); pte 71 arch/x86/include/asm/pgalloc.h pmd_t *pmd, pte_t *pte) pte 73 arch/x86/include/asm/pgalloc.h paravirt_alloc_pte(mm, __pa(pte) >> PAGE_SHIFT); pte 74 arch/x86/include/asm/pgalloc.h set_pmd_safe(pmd, __pmd(__pa(pte) | _PAGE_TABLE)); pte 78 arch/x86/include/asm/pgalloc.h struct page *pte) pte 80 arch/x86/include/asm/pgalloc.h unsigned long pfn = page_to_pfn(pte); pte 15 arch/x86/include/asm/pgtable-2level.h static inline void native_set_pte(pte_t *ptep , pte_t pte) pte 17 arch/x86/include/asm/pgtable-2level.h *ptep = pte; pte 29 arch/x86/include/asm/pgtable-2level.h static inline void native_set_pte_atomic(pte_t *ptep, pte_t pte) pte 31 arch/x86/include/asm/pgtable-2level.h native_set_pte(ptep, pte); pte 95 arch/x86/include/asm/pgtable-2level.h #define __pte_to_swp_entry(pte) ((swp_entry_t) { (pte).pte_low }) pte 96 arch/x86/include/asm/pgtable-2level.h #define __swp_entry_to_pte(x) ((pte_t) { .pte = (x).val }) pte 16 arch/x86/include/asm/pgtable-2level_types.h pteval_t pte; pte 30 arch/x86/include/asm/pgtable-3level.h static inline void native_set_pte(pte_t *ptep, pte_t pte) pte 32 arch/x86/include/asm/pgtable-3level.h ptep->pte_high = pte.pte_high; pte 34 arch/x86/include/asm/pgtable-3level.h ptep->pte_low = pte.pte_low; pte 91 arch/x86/include/asm/pgtable-3level.h static inline void native_set_pte_atomic(pte_t *ptep, pte_t pte) pte 93 arch/x86/include/asm/pgtable-3level.h set_64bit((unsigned long long *)(ptep), native_pte_val(pte)); pte 155 arch/x86/include/asm/pgtable-3level.h res.pte = (pteval_t)arch_atomic64_xchg((atomic64_t *)ptep, 0); pte 274 arch/x86/include/asm/pgtable-3level.h #define __swp_entry_to_pte(x) ((pte_t){ .pte = \ pte 282 arch/x86/include/asm/pgtable-3level.h #define __pteval_swp_type(x) ((unsigned long)((x).pte >> (64 - SWP_TYPE_BITS))) pte 283 arch/x86/include/asm/pgtable-3level.h #define __pteval_swp_offset(x) ((unsigned long)(~((x).pte) << SWP_TYPE_BITS >> SWP_OFFSET_SHIFT)) pte 285 arch/x86/include/asm/pgtable-3level.h #define __pte_to_swp_entry(pte) (__swp_entry(__pteval_swp_type(pte), \ pte 286 arch/x86/include/asm/pgtable-3level.h __pteval_swp_offset(pte))) pte 19 arch/x86/include/asm/pgtable-3level_types.h pteval_t pte; pte 63 arch/x86/include/asm/pgtable.h #define set_pte(ptep, pte) native_set_pte(ptep, pte) pte 64 arch/x86/include/asm/pgtable.h #define set_pte_at(mm, addr, ptep, pte) native_set_pte_at(mm, addr, ptep, pte) pte 66 arch/x86/include/asm/pgtable.h #define set_pte_atomic(ptep, pte) \ pte 67 arch/x86/include/asm/pgtable.h native_set_pte_atomic(ptep, pte) pte 123 arch/x86/include/asm/pgtable.h static inline int pte_dirty(pte_t pte) pte 125 arch/x86/include/asm/pgtable.h return pte_flags(pte) & _PAGE_DIRTY; pte 157 arch/x86/include/asm/pgtable.h static inline int pte_young(pte_t pte) pte 159 arch/x86/include/asm/pgtable.h return pte_flags(pte) & _PAGE_ACCESSED; pte 182 arch/x86/include/asm/pgtable.h static inline int pte_write(pte_t pte) pte 184 arch/x86/include/asm/pgtable.h return pte_flags(pte) & _PAGE_RW; pte 187 arch/x86/include/asm/pgtable.h static inline int pte_huge(pte_t pte) pte 189 arch/x86/include/asm/pgtable.h return pte_flags(pte) & _PAGE_PSE; pte 192 arch/x86/include/asm/pgtable.h static inline int pte_global(pte_t pte) pte 194 arch/x86/include/asm/pgtable.h return pte_flags(pte) & _PAGE_GLOBAL; pte 197 arch/x86/include/asm/pgtable.h static inline int pte_exec(pte_t pte) pte 199 arch/x86/include/asm/pgtable.h return !(pte_flags(pte) & _PAGE_NX); pte 202 arch/x86/include/asm/pgtable.h static inline int pte_special(pte_t pte) pte 204 arch/x86/include/asm/pgtable.h return pte_flags(pte) & _PAGE_SPECIAL; pte 211 arch/x86/include/asm/pgtable.h static inline unsigned long pte_pfn(pte_t pte) pte 213 arch/x86/include/asm/pgtable.h phys_addr_t pfn = pte_val(pte); pte 248 arch/x86/include/asm/pgtable.h #define pte_page(pte) pfn_to_page(pte_pfn(pte)) pte 250 arch/x86/include/asm/pgtable.h static inline int pmd_large(pmd_t pte) pte 252 arch/x86/include/asm/pgtable.h return pmd_flags(pte) & _PAGE_PSE; pte 300 arch/x86/include/asm/pgtable.h static inline pte_t pte_set_flags(pte_t pte, pteval_t set) pte 302 arch/x86/include/asm/pgtable.h pteval_t v = native_pte_val(pte); pte 307 arch/x86/include/asm/pgtable.h static inline pte_t pte_clear_flags(pte_t pte, pteval_t clear) pte 309 arch/x86/include/asm/pgtable.h pteval_t v = native_pte_val(pte); pte 314 arch/x86/include/asm/pgtable.h static inline pte_t pte_mkclean(pte_t pte) pte 316 arch/x86/include/asm/pgtable.h return pte_clear_flags(pte, _PAGE_DIRTY); pte 319 arch/x86/include/asm/pgtable.h static inline pte_t pte_mkold(pte_t pte) pte 321 arch/x86/include/asm/pgtable.h return pte_clear_flags(pte, _PAGE_ACCESSED); pte 324 arch/x86/include/asm/pgtable.h static inline pte_t pte_wrprotect(pte_t pte) pte 326 arch/x86/include/asm/pgtable.h return pte_clear_flags(pte, _PAGE_RW); pte 329 arch/x86/include/asm/pgtable.h static inline pte_t pte_mkexec(pte_t pte) pte 331 arch/x86/include/asm/pgtable.h return pte_clear_flags(pte, _PAGE_NX); pte 334 arch/x86/include/asm/pgtable.h static inline pte_t pte_mkdirty(pte_t pte) pte 336 arch/x86/include/asm/pgtable.h return pte_set_flags(pte, _PAGE_DIRTY | _PAGE_SOFT_DIRTY); pte 339 arch/x86/include/asm/pgtable.h static inline pte_t pte_mkyoung(pte_t pte) pte 341 arch/x86/include/asm/pgtable.h return pte_set_flags(pte, _PAGE_ACCESSED); pte 344 arch/x86/include/asm/pgtable.h static inline pte_t pte_mkwrite(pte_t pte) pte 346 arch/x86/include/asm/pgtable.h return pte_set_flags(pte, _PAGE_RW); pte 349 arch/x86/include/asm/pgtable.h static inline pte_t pte_mkhuge(pte_t pte) pte 351 arch/x86/include/asm/pgtable.h return pte_set_flags(pte, _PAGE_PSE); pte 354 arch/x86/include/asm/pgtable.h static inline pte_t pte_clrhuge(pte_t pte) pte 356 arch/x86/include/asm/pgtable.h return pte_clear_flags(pte, _PAGE_PSE); pte 359 arch/x86/include/asm/pgtable.h static inline pte_t pte_mkglobal(pte_t pte) pte 361 arch/x86/include/asm/pgtable.h return pte_set_flags(pte, _PAGE_GLOBAL); pte 364 arch/x86/include/asm/pgtable.h static inline pte_t pte_clrglobal(pte_t pte) pte 366 arch/x86/include/asm/pgtable.h return pte_clear_flags(pte, _PAGE_GLOBAL); pte 369 arch/x86/include/asm/pgtable.h static inline pte_t pte_mkspecial(pte_t pte) pte 371 arch/x86/include/asm/pgtable.h return pte_set_flags(pte, _PAGE_SPECIAL); pte 374 arch/x86/include/asm/pgtable.h static inline pte_t pte_mkdevmap(pte_t pte) pte 376 arch/x86/include/asm/pgtable.h return pte_set_flags(pte, _PAGE_SPECIAL|_PAGE_DEVMAP); pte 488 arch/x86/include/asm/pgtable.h static inline int pte_soft_dirty(pte_t pte) pte 490 arch/x86/include/asm/pgtable.h return pte_flags(pte) & _PAGE_SOFT_DIRTY; pte 503 arch/x86/include/asm/pgtable.h static inline pte_t pte_mksoft_dirty(pte_t pte) pte 505 arch/x86/include/asm/pgtable.h return pte_set_flags(pte, _PAGE_SOFT_DIRTY); pte 518 arch/x86/include/asm/pgtable.h static inline pte_t pte_clear_soft_dirty(pte_t pte) pte 520 arch/x86/include/asm/pgtable.h return pte_clear_flags(pte, _PAGE_SOFT_DIRTY); pte 604 arch/x86/include/asm/pgtable.h static inline pte_t pte_modify(pte_t pte, pgprot_t newprot) pte 606 arch/x86/include/asm/pgtable.h pteval_t val = pte_val(pte), oldval = val; pte 723 arch/x86/include/asm/pgtable.h static inline int pte_none(pte_t pte) pte 725 arch/x86/include/asm/pgtable.h return !(pte.pte & ~(_PAGE_KNL_ERRATUM_MASK)); pte 731 arch/x86/include/asm/pgtable.h return a.pte == b.pte; pte 775 arch/x86/include/asm/pgtable.h static inline int pte_protnone(pte_t pte) pte 777 arch/x86/include/asm/pgtable.h return (pte_flags(pte) & (_PAGE_PROTNONE | _PAGE_PRESENT)) pte 1084 arch/x86/include/asm/pgtable.h pte_t *ptep , pte_t pte) pte 1086 arch/x86/include/asm/pgtable.h native_set_pte(ptep, pte); pte 1127 arch/x86/include/asm/pgtable.h pte_t pte = native_ptep_get_and_clear(ptep); pte 1128 arch/x86/include/asm/pgtable.h return pte; pte 1136 arch/x86/include/asm/pgtable.h pte_t pte; pte 1142 arch/x86/include/asm/pgtable.h pte = native_local_ptep_get_and_clear(ptep); pte 1144 arch/x86/include/asm/pgtable.h pte = ptep_get_and_clear(mm, addr, ptep); pte 1146 arch/x86/include/asm/pgtable.h return pte; pte 1153 arch/x86/include/asm/pgtable.h clear_bit(_PAGE_BIT_RW, (unsigned long *)&ptep->pte); pte 1345 arch/x86/include/asm/pgtable.h static inline pte_t pte_swp_mksoft_dirty(pte_t pte) pte 1347 arch/x86/include/asm/pgtable.h return pte_set_flags(pte, _PAGE_SWP_SOFT_DIRTY); pte 1350 arch/x86/include/asm/pgtable.h static inline int pte_swp_soft_dirty(pte_t pte) pte 1352 arch/x86/include/asm/pgtable.h return pte_flags(pte) & _PAGE_SWP_SOFT_DIRTY; pte 1355 arch/x86/include/asm/pgtable.h static inline pte_t pte_swp_clear_soft_dirty(pte_t pte) pte 1357 arch/x86/include/asm/pgtable.h return pte_clear_flags(pte, _PAGE_SWP_SOFT_DIRTY); pte 1445 arch/x86/include/asm/pgtable.h static inline bool pte_access_permitted(pte_t pte, bool write) pte 1447 arch/x86/include/asm/pgtable.h return __pte_access_permitted(pte_val(pte), write); pte 52 arch/x86/include/asm/pgtable_32.h #define pte_unmap(pte) kunmap_atomic((pte)) pte 56 arch/x86/include/asm/pgtable_32.h #define pte_unmap(pte) do { } while (0) pte 59 arch/x86/include/asm/pgtable_64.h static inline void native_set_pte(pte_t *ptep, pte_t pte) pte 61 arch/x86/include/asm/pgtable_64.h WRITE_ONCE(*ptep, pte); pte 70 arch/x86/include/asm/pgtable_64.h static inline void native_set_pte_atomic(pte_t *ptep, pte_t pte) pte 72 arch/x86/include/asm/pgtable_64.h native_set_pte(ptep, pte); pte 88 arch/x86/include/asm/pgtable_64.h return native_make_pte(xchg(&xp->pte, 0)); pte 185 arch/x86/include/asm/pgtable_64.h #define pte_unmap(pte) ((void)(pte))/* NOP */ pte 233 arch/x86/include/asm/pgtable_64.h #define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val((pte)) }) pte 235 arch/x86/include/asm/pgtable_64.h #define __swp_entry_to_pte(x) ((pte_t) { .pte = (x).val }) pte 21 arch/x86/include/asm/pgtable_64_types.h typedef struct { pteval_t pte; } pte_t; pte 439 arch/x86/include/asm/pgtable_types.h return (pte_t) { .pte = val }; pte 442 arch/x86/include/asm/pgtable_types.h static inline pteval_t native_pte_val(pte_t pte) pte 444 arch/x86/include/asm/pgtable_types.h return pte.pte; pte 447 arch/x86/include/asm/pgtable_types.h static inline pteval_t pte_flags(pte_t pte) pte 449 arch/x86/include/asm/pgtable_types.h return native_pte_val(pte) & PTE_FLAGS_MASK; pte 529 arch/x86/include/asm/pgtable_types.h void set_pte_vaddr(unsigned long vaddr, pte_t pte); pte 349 arch/x86/include/asm/xen/hypercall.h new_val.pte, flags); pte 352 arch/x86/include/asm/xen/hypercall.h new_val.pte, new_val.pte >> 32, flags); pte 465 arch/x86/include/asm/xen/hypercall.h mcl->args[1] = new_val.pte; pte 468 arch/x86/include/asm/xen/hypercall.h mcl->args[1] = new_val.pte; pte 469 arch/x86/include/asm/xen/hypercall.h mcl->args[2] = new_val.pte >> 32; pte 315 arch/x86/include/asm/xen/page.h static inline unsigned long pte_mfn(pte_t pte) pte 317 arch/x86/include/asm/xen/page.h return (pte.pte & XEN_PTE_MFN_MASK) >> PAGE_SHIFT; pte 322 arch/x86/include/asm/xen/page.h pte_t pte; pte 324 arch/x86/include/asm/xen/page.h pte.pte = ((phys_addr_t)page_nr << PAGE_SHIFT) | pte 327 arch/x86/include/asm/xen/page.h return pte; pte 330 arch/x86/include/asm/xen/page.h static inline pteval_t pte_val_ma(pte_t pte) pte 332 arch/x86/include/asm/xen/page.h return pte.pte; pte 337 arch/x86/include/asm/xen/page.h return (pte_t) { .pte = x }; pte 794 arch/x86/kernel/alternative.c pte_t pte, *ptep; pte 838 arch/x86/kernel/alternative.c pte = mk_pte(pages[0], pgprot); pte 839 arch/x86/kernel/alternative.c set_pte_at(poking_mm, poking_addr, ptep, pte); pte 842 arch/x86/kernel/alternative.c pte = mk_pte(pages[1], pgprot); pte 843 arch/x86/kernel/alternative.c set_pte_at(poking_mm, poking_addr + PAGE_SIZE, ptep + 1, pte); pte 137 arch/x86/kernel/espfix_64.c pte_t pte, *pte_p; pte 194 arch/x86/kernel/espfix_64.c pte = __pte(__pa(stack_page) | ((__PAGE_KERNEL_RO | _PAGE_ENC) & ptemask)); pte 196 arch/x86/kernel/espfix_64.c set_pte(&pte_p[n*PTE_STRIDE], pte); pte 78 arch/x86/kernel/head32.c pte_t pte, *ptep; pte 93 arch/x86/kernel/head32.c pte.pte = PTE_IDENT_ATTR; pte 95 arch/x86/kernel/head32.c while ((pte.pte & PTE_PFN_MASK) < limit) { pte 104 arch/x86/kernel/head32.c *ptep = pte; pte 105 arch/x86/kernel/head32.c pte.pte += PAGE_SIZE; pte 114 arch/x86/kernel/head32.c *ptr = (pte.pte & PTE_PFN_MASK) >> PAGE_SHIFT; pte 232 arch/x86/kernel/ldt.c pte_t pte, *ptep; pte 253 arch/x86/kernel/ldt.c pte = pfn_pte(pfn, pte_prot); pte 254 arch/x86/kernel/ldt.c set_pte_at(mm, va, ptep, pte); pte 92 arch/x86/kernel/machine_kexec_32.c pgd_t *pgd, pmd_t *pmd, pte_t *pte, pte 107 arch/x86/kernel/machine_kexec_32.c set_pmd(pmd, __pmd(__pa(pte) | _PAGE_TABLE)); pte 108 arch/x86/kernel/machine_kexec_32.c pte = pte_offset_kernel(pmd, vaddr); pte 109 arch/x86/kernel/machine_kexec_32.c set_pte(pte, pfn_pte(paddr >> PAGE_SHIFT, PAGE_KERNEL_EXEC)); pte 120 arch/x86/kernel/machine_kexec_64.c free_page((unsigned long)image->arch.pte); pte 121 arch/x86/kernel/machine_kexec_64.c image->arch.pte = NULL; pte 132 arch/x86/kernel/machine_kexec_64.c pte_t *pte; pte 162 arch/x86/kernel/machine_kexec_64.c pte = (pte_t *)get_zeroed_page(GFP_KERNEL); pte 163 arch/x86/kernel/machine_kexec_64.c if (!pte) pte 165 arch/x86/kernel/machine_kexec_64.c image->arch.pte = pte; pte 166 arch/x86/kernel/machine_kexec_64.c set_pmd(pmd, __pmd(__pa(pte) | _KERNPG_TABLE)); pte 168 arch/x86/kernel/machine_kexec_64.c pte = pte_offset_kernel(pmd, vaddr); pte 173 arch/x86/kernel/machine_kexec_64.c set_pte(pte, pfn_pte(paddr >> PAGE_SHIFT, prot)); pte 111 arch/x86/kernel/tboot.c pte_t *pte; pte 123 arch/x86/kernel/tboot.c pte = pte_alloc_map(&tboot_mm, pmd, vaddr); pte 124 arch/x86/kernel/tboot.c if (!pte) pte 126 arch/x86/kernel/tboot.c set_pte_at(&tboot_mm, vaddr, pte, pfn_pte(pfn, prot)); pte 127 arch/x86/kernel/tboot.c pte_unmap(pte); pte 172 arch/x86/kernel/vm86_32.c pte_t *pte; pte 193 arch/x86/kernel/vm86_32.c pte = pte_offset_map_lock(mm, pmd, 0xA0000, &ptl); pte 195 arch/x86/kernel/vm86_32.c if (pte_present(*pte)) pte 196 arch/x86/kernel/vm86_32.c set_pte(pte, pte_wrprotect(*pte)); pte 197 arch/x86/kernel/vm86_32.c pte++; pte 199 arch/x86/kernel/vm86_32.c pte_unmap_unlock(pte, ptl); pte 609 arch/x86/kvm/mmu.c static int is_shadow_present_pte(u64 pte) pte 611 arch/x86/kvm/mmu.c return (pte != 0) && !is_mmio_spte(pte); pte 614 arch/x86/kvm/mmu.c static int is_large_pte(u64 pte) pte 616 arch/x86/kvm/mmu.c return pte & PT_PAGE_SIZE_MASK; pte 619 arch/x86/kvm/mmu.c static int is_last_spte(u64 pte, int level) pte 623 arch/x86/kvm/mmu.c if (is_large_pte(pte)) pte 633 arch/x86/kvm/mmu.c static kvm_pfn_t spte_to_pfn(u64 pte) pte 635 arch/x86/kvm/mmu.c return (pte & PT64_BASE_ADDR_MASK) >> PAGE_SHIFT; pte 2053 arch/x86/kvm/mmu.c int kvm_set_spte_hva(struct kvm *kvm, unsigned long hva, pte_t pte) pte 2055 arch/x86/kvm/mmu.c return kvm_handle_hva(kvm, hva, (unsigned long)&pte, kvm_set_pte_rmapp); pte 2236 arch/x86/kvm/mmu.c const void *pte) pte 2751 arch/x86/kvm/mmu.c u64 pte; pte 2754 arch/x86/kvm/mmu.c pte = *spte; pte 2755 arch/x86/kvm/mmu.c if (is_shadow_present_pte(pte)) { pte 2756 arch/x86/kvm/mmu.c if (is_last_spte(pte, sp->role.level)) { pte 2758 arch/x86/kvm/mmu.c if (is_large_pte(pte)) pte 2761 arch/x86/kvm/mmu.c child = page_header(pte & PT64_BASE_ADDR_MASK); pte 2767 arch/x86/kvm/mmu.c if (is_mmio_spte(pte)) pte 3174 arch/x86/kvm/mmu.c u64 pte = *sptep; pte 3176 arch/x86/kvm/mmu.c child = page_header(pte & PT64_BASE_ADDR_MASK); pte 4008 arch/x86/kvm/mmu.c __is_rsvd_bits_set(struct rsvd_bits_validate *rsvd_check, u64 pte, int level) pte 4010 arch/x86/kvm/mmu.c int bit7 = (pte >> 7) & 1, low6 = pte & 0x3f; pte 4012 arch/x86/kvm/mmu.c return (pte & rsvd_check->rsvd_bits_mask[bit7][level-1]) | pte 138 arch/x86/kvm/mmu.h static inline int is_writable_pte(unsigned long pte) pte 140 arch/x86/kvm/mmu.h return pte & PT_WRITABLE_MASK; pte 83 arch/x86/kvm/mmutrace.h TP_PROTO(u64 pte, int level), pte 84 arch/x86/kvm/mmutrace.h TP_ARGS(pte, level), pte 87 arch/x86/kvm/mmutrace.h __field(__u64, pte) pte 92 arch/x86/kvm/mmutrace.h __entry->pte = pte; pte 96 arch/x86/kvm/mmutrace.h TP_printk("pte %llx level %u", __entry->pte, __entry->level) pte 78 arch/x86/kvm/paging_tmpl.h #define gpte_to_gfn(pte) gpte_to_gfn_lvl((pte), PT_PAGE_TABLE_LEVEL) pte 122 arch/x86/kvm/paging_tmpl.h static inline int FNAME(is_present_gpte)(unsigned long pte) pte 125 arch/x86/kvm/paging_tmpl.h return pte & PT_PRESENT_MASK; pte 127 arch/x86/kvm/paging_tmpl.h return pte & 7; pte 226 arch/x86/kvm/paging_tmpl.h pt_element_t pte, orig_pte; pte 236 arch/x86/kvm/paging_tmpl.h pte = orig_pte = walker->ptes[level - 1]; pte 240 arch/x86/kvm/paging_tmpl.h if (!(pte & PT_GUEST_ACCESSED_MASK)) { pte 241 arch/x86/kvm/paging_tmpl.h trace_kvm_mmu_set_accessed_bit(table_gfn, index, sizeof(pte)); pte 242 arch/x86/kvm/paging_tmpl.h pte |= PT_GUEST_ACCESSED_MASK; pte 245 arch/x86/kvm/paging_tmpl.h !(pte & PT_GUEST_DIRTY_MASK)) { pte 246 arch/x86/kvm/paging_tmpl.h trace_kvm_mmu_set_dirty_bit(table_gfn, index, sizeof(pte)); pte 251 arch/x86/kvm/paging_tmpl.h pte |= PT_GUEST_DIRTY_MASK; pte 253 arch/x86/kvm/paging_tmpl.h if (pte == orig_pte) pte 272 arch/x86/kvm/paging_tmpl.h ret = FNAME(cmpxchg_gpte)(vcpu, mmu, ptep_user, index, orig_pte, pte); pte 277 arch/x86/kvm/paging_tmpl.h walker->ptes[level - 1] = pte; pte 286 arch/x86/kvm/paging_tmpl.h pte_t pte = {.pte = gpte}; pte 288 arch/x86/kvm/paging_tmpl.h pkeys = pte_flags_pkey(pte_flags(pte)); pte 301 arch/x86/kvm/paging_tmpl.h pt_element_t pte; pte 321 arch/x86/kvm/paging_tmpl.h pte = mmu->get_cr3(vcpu); pte 327 arch/x86/kvm/paging_tmpl.h pte = mmu->get_pdptr(vcpu, (addr >> 30) & 3); pte 328 arch/x86/kvm/paging_tmpl.h trace_kvm_mmu_paging_element(pte, walker->level); pte 329 arch/x86/kvm/paging_tmpl.h if (!FNAME(is_present_gpte)(pte)) pte 355 arch/x86/kvm/paging_tmpl.h table_gfn = gpte_to_gfn(pte); pte 388 arch/x86/kvm/paging_tmpl.h if (unlikely(__copy_from_user(&pte, ptep_user, sizeof(pte)))) pte 392 arch/x86/kvm/paging_tmpl.h trace_kvm_mmu_paging_element(pte, walker->level); pte 398 arch/x86/kvm/paging_tmpl.h pte_access = pt_access & (pte ^ walk_nx_mask); pte 400 arch/x86/kvm/paging_tmpl.h if (unlikely(!FNAME(is_present_gpte)(pte))) pte 403 arch/x86/kvm/paging_tmpl.h if (unlikely(is_rsvd_bits_set(mmu, pte, walker->level))) { pte 408 arch/x86/kvm/paging_tmpl.h walker->ptes[walker->level - 1] = pte; pte 409 arch/x86/kvm/paging_tmpl.h } while (!is_last_gpte(mmu, walker->level, pte)); pte 411 arch/x86/kvm/paging_tmpl.h pte_pkey = FNAME(gpte_pkeys)(vcpu, pte); pte 421 arch/x86/kvm/paging_tmpl.h gfn = gpte_to_gfn_lvl(pte, walker->level); pte 425 arch/x86/kvm/paging_tmpl.h gfn += pse36_gfn_delta(pte); pte 434 arch/x86/kvm/paging_tmpl.h FNAME(protect_clean_gpte)(mmu, &walker->pte_access, pte); pte 441 arch/x86/kvm/paging_tmpl.h accessed_dirty &= pte >> pte 453 arch/x86/kvm/paging_tmpl.h __func__, (u64)pte, walker->pte_access, walker->pt_access); pte 548 arch/x86/kvm/paging_tmpl.h u64 *spte, const void *pte) pte 550 arch/x86/kvm/paging_tmpl.h pt_element_t gpte = *(const pt_element_t *)pte; pte 32 arch/x86/mm/cpu_entry_area.c pte_t pte = pfn_pte(pa >> PAGE_SHIFT, flags); pte 43 arch/x86/mm/cpu_entry_area.c pte = pte_set_flags(pte, _PAGE_GLOBAL); pte 45 arch/x86/mm/cpu_entry_area.c set_pte_vaddr(va, pte); pte 361 arch/x86/mm/dump_pagetables.c pte_t *pte; pte 366 arch/x86/mm/dump_pagetables.c pte = pte_offset_map(&addr, st->current_address); pte 367 arch/x86/mm/dump_pagetables.c prot = pte_flags(*pte); pte 370 arch/x86/mm/dump_pagetables.c pte_unmap(pte); pte 298 arch/x86/mm/fault.c pte_t *pte; pte 323 arch/x86/mm/fault.c pte = pte_offset_kernel(pmd, address); pte 324 arch/x86/mm/fault.c pr_cont("*pte = %0*Lx ", sizeof(*pte) * 2, (u64)pte_val(*pte)); pte 359 arch/x86/mm/fault.c pte_t *pte; pte 413 arch/x86/mm/fault.c pte = pte_offset_kernel(pmd, address); pte 414 arch/x86/mm/fault.c if (!pte_present(*pte)) pte 453 arch/x86/mm/fault.c pte_t *pte; pte 487 arch/x86/mm/fault.c pte = pte_offset_kernel(pmd, address); pte 488 arch/x86/mm/fault.c if (bad_address(pte)) pte 491 arch/x86/mm/fault.c pr_cont("PTE %lx", pte_val(*pte)); pte 616 arch/x86/mm/fault.c pte_t *pte; pte 621 arch/x86/mm/fault.c pte = lookup_address_in_pgd(pgd, address, &level); pte 623 arch/x86/mm/fault.c if (pte && pte_present(*pte) && !pte_exec(*pte)) pte 626 arch/x86/mm/fault.c if (pte && pte_present(*pte) && pte_exec(*pte) && pte 1088 arch/x86/mm/fault.c static int spurious_kernel_fault_check(unsigned long error_code, pte_t *pte) pte 1090 arch/x86/mm/fault.c if ((error_code & X86_PF_WRITE) && !pte_write(*pte)) pte 1093 arch/x86/mm/fault.c if ((error_code & X86_PF_INSTR) && !pte_exec(*pte)) pte 1127 arch/x86/mm/fault.c pte_t *pte; pte 1168 arch/x86/mm/fault.c pte = pte_offset_kernel(pmd, address); pte 1169 arch/x86/mm/fault.c if (!pte_present(*pte)) pte 1172 arch/x86/mm/fault.c ret = spurious_kernel_fault_check(error_code, pte); pte 38 arch/x86/mm/hugetlbpage.c pte = huge_pte_offset(mm, address, vma_mmu_pagesize(vma)); pte 41 arch/x86/mm/hugetlbpage.c WARN_ON(!pte || pte_none(*pte)); pte 43 arch/x86/mm/hugetlbpage.c page = &pte_page(*pte)[vpfn % (HPAGE_SIZE/PAGE_SIZE)]; pte 157 arch/x86/mm/init_32.c static pte_t *__init page_table_kmap_check(pte_t *pte, pmd_t *pmd, pte 180 arch/x86/mm/init_32.c set_pte(newpte + i, pte[i]); pte 188 arch/x86/mm/init_32.c paravirt_release_pte(__pa(pte) >> PAGE_SHIFT); pte 189 arch/x86/mm/init_32.c pte = newpte; pte 193 arch/x86/mm/init_32.c && lastpte && lastpte + PTRS_PER_PTE != pte); pte 195 arch/x86/mm/init_32.c return pte; pte 214 arch/x86/mm/init_32.c pte_t *pte = NULL; pte 231 arch/x86/mm/init_32.c pte = page_table_kmap_check(one_page_table_init(pmd), pte 232 arch/x86/mm/init_32.c pmd, vaddr, pte, &adr); pte 265 arch/x86/mm/init_32.c pte_t *pte; pte 343 arch/x86/mm/init_32.c pte = one_page_table_init(pmd); pte 346 arch/x86/mm/init_32.c pte += pte_ofs; pte 348 arch/x86/mm/init_32.c pte++, pfn++, pte_ofs++, addr += PAGE_SIZE) { pte 361 arch/x86/mm/init_32.c set_pte(pte, pfn_pte(pfn, init_prot)); pte 364 arch/x86/mm/init_32.c set_pte(pte, pfn_pte(pfn, prot)); pte 421 arch/x86/mm/init_32.c pte_t *pte; pte 430 arch/x86/mm/init_32.c pte = pte_offset_kernel(pmd, vaddr); pte 431 arch/x86/mm/init_32.c pkmap_page_table = pte; pte 478 arch/x86/mm/init_32.c pte_t *pte; pte 508 arch/x86/mm/init_32.c pte = pte_offset_kernel(pmd, va); pte 509 arch/x86/mm/init_32.c if (!pte_present(*pte)) pte 513 arch/x86/mm/init_32.c pfn, pmd, __pa(pmd), pte, __pa(pte)); pte 514 arch/x86/mm/init_32.c pte_clear(NULL, va, pte); pte 75 arch/x86/mm/init_64.c DEFINE_POPULATE(pmd_populate_kernel, pmd, pte, init) pte 90 arch/x86/mm/init_64.c DEFINE_ENTRY(pte, pte, init) pte 282 arch/x86/mm/init_64.c pte_t *pte = (pte_t *) spp_getpage(); pte 283 arch/x86/mm/init_64.c pmd_populate_kernel(&init_mm, pmd, pte); pte 284 arch/x86/mm/init_64.c if (pte != pte_offset_kernel(pmd, 0)) pte 293 arch/x86/mm/init_64.c pte_t *pte = fill_pte(pmd, vaddr); pte 295 arch/x86/mm/init_64.c set_pte(pte, new_pte); pte 453 arch/x86/mm/init_64.c pte_t *pte; pte 456 arch/x86/mm/init_64.c pte = pte_page + pte_index(paddr); pte 459 arch/x86/mm/init_64.c for (; i < PTRS_PER_PTE; i++, paddr = paddr_next, pte++) { pte 467 arch/x86/mm/init_64.c set_pte_init(pte, __pte(0), init); pte 477 arch/x86/mm/init_64.c if (!pte_none(*pte)) { pte 484 arch/x86/mm/init_64.c pr_info(" pte=%p addr=%lx pte=%016lx\n", pte, paddr, pte 485 arch/x86/mm/init_64.c pfn_pte(paddr >> PAGE_SHIFT, PAGE_KERNEL).pte); pte 487 arch/x86/mm/init_64.c set_pte_init(pte, pfn_pte(paddr >> PAGE_SHIFT, prot), init); pte 512 arch/x86/mm/init_64.c pte_t *pte; pte 529 arch/x86/mm/init_64.c pte = (pte_t *)pmd_page_vaddr(*pmd); pte 530 arch/x86/mm/init_64.c paddr_last = phys_pte_init(pte, paddr, pte 569 arch/x86/mm/init_64.c pte = alloc_low_page(); pte 570 arch/x86/mm/init_64.c paddr_last = phys_pte_init(pte, paddr, paddr_end, new_prot, init); pte 573 arch/x86/mm/init_64.c pmd_populate_kernel_init(&init_mm, pmd, pte, init); pte 905 arch/x86/mm/init_64.c pte_t *pte; pte 909 arch/x86/mm/init_64.c pte = pte_start + i; pte 910 arch/x86/mm/init_64.c if (!pte_none(*pte)) pte 962 arch/x86/mm/init_64.c pte_t *pte; pte 966 arch/x86/mm/init_64.c pte = pte_start + pte_index(addr); pte 967 arch/x86/mm/init_64.c for (; addr < end; addr = next, pte++) { pte 972 arch/x86/mm/init_64.c if (!pte_present(*pte)) pte 980 arch/x86/mm/init_64.c phys_addr = pte_val(*pte) + (addr & PAGE_MASK); pte 990 arch/x86/mm/init_64.c free_pagetable(pte_page(*pte), 0); pte 993 arch/x86/mm/init_64.c pte_clear(&init_mm, addr, pte); pte 1011 arch/x86/mm/init_64.c page_addr = page_address(pte_page(*pte)); pte 1013 arch/x86/mm/init_64.c free_pagetable(pte_page(*pte), 0); pte 1016 arch/x86/mm/init_64.c pte_clear(&init_mm, addr, pte); pte 1348 arch/x86/mm/init_64.c pte_t *pte; pte 1375 arch/x86/mm/init_64.c pte = pte_offset_kernel(pmd, addr); pte 1376 arch/x86/mm/init_64.c if (pte_none(*pte)) pte 1379 arch/x86/mm/init_64.c return pfn_valid(pte_pfn(*pte)); pte 1549 arch/x86/mm/init_64.c pte_t *pte = NULL; pte 1580 arch/x86/mm/init_64.c pte = pte_offset_kernel(pmd, addr); pte 1581 arch/x86/mm/init_64.c if (pte_none(*pte)) pte 1583 arch/x86/mm/init_64.c get_page_bootmem(section_nr, pte_page(*pte), pte 866 arch/x86/mm/ioremap.c pte_t *pte; pte 872 arch/x86/mm/ioremap.c pte = early_ioremap_pte(addr); pte 878 arch/x86/mm/ioremap.c set_pte(pte, pfn_pte(phys >> PAGE_SHIFT, flags)); pte 880 arch/x86/mm/ioremap.c pte_clear(&init_mm, addr, pte); pte 42 arch/x86/mm/kasan_init_64.c pte_t *pte; pte 61 arch/x86/mm/kasan_init_64.c pte = pte_offset_kernel(pmd, addr); pte 66 arch/x86/mm/kasan_init_64.c if (!pte_none(*pte)) pte 71 arch/x86/mm/kasan_init_64.c set_pte_at(&init_mm, addr, pte, entry); pte 72 arch/x86/mm/kasan_init_64.c } while (pte++, addr += PAGE_SIZE, addr != end); pte 382 arch/x86/mm/kasan_init_64.c pte_t pte; pte 388 arch/x86/mm/kasan_init_64.c pte = __pte(__pa(kasan_early_shadow_page) | pgprot_val(prot)); pte 389 arch/x86/mm/kasan_init_64.c set_pte(&kasan_early_shadow_pte[i], pte); pte 77 arch/x86/mm/kmmio.c pte_t *pte = lookup_address(addr, &l); pte 79 arch/x86/mm/kmmio.c if (!pte) pte 114 arch/x86/mm/kmmio.c pte_t *pte = lookup_address(addr, &l); pte 116 arch/x86/mm/kmmio.c if (!pte) pte 141 arch/x86/mm/kmmio.c static void clear_pte_presence(pte_t *pte, bool clear, pteval_t *old) pte 143 arch/x86/mm/kmmio.c pteval_t v = pte_val(*pte); pte 147 arch/x86/mm/kmmio.c pte_clear(&init_mm, 0, pte); pte 150 arch/x86/mm/kmmio.c set_pte_atomic(pte, __pte(*old)); pte 157 arch/x86/mm/kmmio.c pte_t *pte = lookup_address(f->addr, &level); pte 159 arch/x86/mm/kmmio.c if (!pte) { pte 166 arch/x86/mm/kmmio.c clear_pmd_presence((pmd_t *)pte, clear, &f->old_presence); pte 169 arch/x86/mm/kmmio.c clear_pte_presence(pte, clear, &f->old_presence); pte 237 arch/x86/mm/kmmio.c pte_t *pte = lookup_address(addr, &l); pte 238 arch/x86/mm/kmmio.c if (!pte) pte 446 arch/x86/mm/kmmio.c pte_t *pte; pte 454 arch/x86/mm/kmmio.c pte = lookup_address(addr, &l); pte 455 arch/x86/mm/kmmio.c if (!pte) { pte 542 arch/x86/mm/kmmio.c pte_t *pte; pte 544 arch/x86/mm/kmmio.c pte = lookup_address(addr, &l); pte 545 arch/x86/mm/kmmio.c if (!pte) pte 162 arch/x86/mm/mem_encrypt_identity.c pte_t *pte; pte 170 arch/x86/mm/mem_encrypt_identity.c pte = ppd->pgtable_area; pte 171 arch/x86/mm/mem_encrypt_identity.c memset(pte, 0, sizeof(*pte) * PTRS_PER_PTE); pte 172 arch/x86/mm/mem_encrypt_identity.c ppd->pgtable_area += sizeof(*pte) * PTRS_PER_PTE; pte 173 arch/x86/mm/mem_encrypt_identity.c set_pmd(pmd, __pmd(PMD_FLAGS | __pa(pte))); pte 179 arch/x86/mm/mem_encrypt_identity.c pte = pte_offset_map(pmd, ppd->vaddr); pte 180 arch/x86/mm/mem_encrypt_identity.c if (pte_none(*pte)) pte 181 arch/x86/mm/mem_encrypt_identity.c set_pte(pte, __pte(ppd->paddr | ppd->pte_flags)); pte 84 arch/x86/mm/mmio-mod.c pte_t *pte = lookup_address(address, &level); pte 86 arch/x86/mm/mmio-mod.c if (!pte) { pte 99 arch/x86/mm/mmio-mod.c (unsigned long long)pte_val(*pte), pte 100 arch/x86/mm/mmio-mod.c (unsigned long long)pte_val(*pte) & _PAGE_PRESENT); pte 40 arch/x86/mm/pageattr-test.c static int pte_testbit(pte_t pte) pte 42 arch/x86/mm/pageattr-test.c return pte_flags(pte) & _PAGE_SOFTW1; pte 61 arch/x86/mm/pageattr-test.c pte_t *pte; pte 63 arch/x86/mm/pageattr-test.c pte = lookup_address(addr, &level); pte 64 arch/x86/mm/pageattr-test.c if (!pte) { pte 74 arch/x86/mm/pageattr-test.c if ((pte_val(*pte) & _PAGE_PRESENT) && !(pte_val(*pte) & _PAGE_PSE)) { pte 77 arch/x86/mm/pageattr-test.c addr, level, (u64)pte_val(*pte)); pte 86 arch/x86/mm/pageattr-test.c if (!(pte_val(*pte) & _PAGE_NX)) { pte 122 arch/x86/mm/pageattr-test.c pte_t *pte, pte0; pte 149 arch/x86/mm/pageattr-test.c pte = NULL; pte 153 arch/x86/mm/pageattr-test.c pte = lookup_address(addr[i] + k*PAGE_SIZE, &level); pte 154 arch/x86/mm/pageattr-test.c if (!pte || pgprot_val(pte_pgprot(*pte)) == 0 || pte 155 arch/x86/mm/pageattr-test.c !(pte_val(*pte) & _PAGE_PRESENT)) { pte 160 arch/x86/mm/pageattr-test.c pte0 = *pte; pte 162 arch/x86/mm/pageattr-test.c if (pgprot_val(pte_pgprot(*pte)) != pte 176 arch/x86/mm/pageattr-test.c if (!addr[i] || !pte || !k) { pte 201 arch/x86/mm/pageattr-test.c pte = lookup_address(addr[i], &level); pte 202 arch/x86/mm/pageattr-test.c if (!pte || !pte_testbit(*pte) || pte_huge(*pte)) { pte 204 arch/x86/mm/pageattr-test.c pte ? (u64)pte_val(*pte) : 0ULL); pte 221 arch/x86/mm/pageattr-test.c pte = lookup_address(addr[i], &level); pte 222 arch/x86/mm/pageattr-test.c if (!pte) { pte 232 arch/x86/mm/pageattr-test.c pte = lookup_address(addr[i], &level); pte 233 arch/x86/mm/pageattr-test.c if (!pte || pte_testbit(*pte)) { pte 235 arch/x86/mm/pageattr-test.c addr[i], pte ? (u64)pte_val(*pte) : 0ULL); pte 369 arch/x86/mm/pageattr.c pte_t *pte = lookup_address(addr, &level); pte 374 arch/x86/mm/pageattr.c if (pte && (pte_val(*pte) & _PAGE_PRESENT)) pte 674 arch/x86/mm/pageattr.c pte_t *pte; pte 676 arch/x86/mm/pageattr.c pte = lookup_address(virt_addr, &level); pte 677 arch/x86/mm/pageattr.c BUG_ON(!pte); pte 686 arch/x86/mm/pageattr.c phys_addr = (phys_addr_t)pud_pfn(*(pud_t *)pte) << PAGE_SHIFT; pte 690 arch/x86/mm/pageattr.c phys_addr = (phys_addr_t)pmd_pfn(*(pmd_t *)pte) << PAGE_SHIFT; pte 694 arch/x86/mm/pageattr.c phys_addr = (phys_addr_t)pte_pfn(*pte) << PAGE_SHIFT; pte 705 arch/x86/mm/pageattr.c static void __set_pmd_pte(pte_t *kpte, unsigned long address, pte_t pte) pte 708 arch/x86/mm/pageattr.c set_pte_atomic(kpte, pte); pte 723 arch/x86/mm/pageattr.c set_pte_atomic((pte_t *)pmd, pte); pte 904 arch/x86/mm/pageattr.c static void split_set_pte(struct cpa_data *cpa, pte_t *pte, unsigned long pfn, pte 937 arch/x86/mm/pageattr.c set_pte(pte, pfn_pte(pfn, ref_prot)); pte 1064 arch/x86/mm/pageattr.c static bool try_to_free_pte_page(pte_t *pte) pte 1069 arch/x86/mm/pageattr.c if (!pte_none(pte[i])) pte 1072 arch/x86/mm/pageattr.c free_page((unsigned long)pte); pte 1090 arch/x86/mm/pageattr.c pte_t *pte = pte_offset_kernel(pmd, start); pte 1093 arch/x86/mm/pageattr.c set_pte(pte, __pte(0)); pte 1096 arch/x86/mm/pageattr.c pte++; pte 1203 arch/x86/mm/pageattr.c pte_t *pte = (pte_t *)get_zeroed_page(GFP_KERNEL); pte 1204 arch/x86/mm/pageattr.c if (!pte) pte 1207 arch/x86/mm/pageattr.c set_pmd(pmd, __pmd(__pa(pte) | _KERNPG_TABLE)); pte 1225 arch/x86/mm/pageattr.c pte_t *pte; pte 1227 arch/x86/mm/pageattr.c pte = pte_offset_kernel(pmd, start); pte 1232 arch/x86/mm/pageattr.c set_pte(pte, pfn_pte(cpa->pfn, pgprot)); pte 1236 arch/x86/mm/pageattr.c pte++; pte 2201 arch/x86/mm/pageattr.c pte_t *pte; pte 2206 arch/x86/mm/pageattr.c pte = lookup_address((unsigned long)page_address(page), &level); pte 2207 arch/x86/mm/pageattr.c return (pte_val(*pte) & _PAGE_PRESENT); pte 46 arch/x86/mm/pgtable.c void ___pte_free_tlb(struct mmu_gather *tlb, struct page *pte) pte 48 arch/x86/mm/pgtable.c pgtable_pte_page_dtor(pte); pte 49 arch/x86/mm/pgtable.c paravirt_release_pte(page_to_pfn(pte)); pte 50 arch/x86/mm/pgtable.c paravirt_tlb_remove_table(tlb, pte); pte 541 arch/x86/mm/pgtable.c (unsigned long *) &ptep->pte); pte 625 arch/x86/mm/pgtable.c void __native_set_fixmap(enum fixed_addresses idx, pte_t pte) pte 642 arch/x86/mm/pgtable.c set_pte_vaddr(address, pte); pte 802 arch/x86/mm/pgtable.c pte_t *pte; pte 823 arch/x86/mm/pgtable.c pte = (pte_t *)pmd_page_vaddr(pmd_sv[i]); pte 824 arch/x86/mm/pgtable.c free_page((unsigned long)pte); pte 844 arch/x86/mm/pgtable.c pte_t *pte; pte 846 arch/x86/mm/pgtable.c pte = (pte_t *)pmd_page_vaddr(*pmd); pte 852 arch/x86/mm/pgtable.c free_page((unsigned long)pte); pte 34 arch/x86/mm/pgtable_32.c pte_t *pte; pte 56 arch/x86/mm/pgtable_32.c pte = pte_offset_kernel(pmd, vaddr); pte 58 arch/x86/mm/pgtable_32.c set_pte_at(&init_mm, vaddr, pte, pteval); pte 60 arch/x86/mm/pgtable_32.c pte_clear(&init_mm, vaddr, pte); pte 249 arch/x86/mm/pti.c pte_t *pte; pte 269 arch/x86/mm/pti.c pte = pte_offset_kernel(pmd, address); pte 270 arch/x86/mm/pti.c if (pte_flags(*pte) & _PAGE_USER) { pte 274 arch/x86/mm/pti.c return pte; pte 280 arch/x86/mm/pti.c pte_t *pte, *target_pte; pte 283 arch/x86/mm/pti.c pte = lookup_address(VSYSCALL_ADDR, &level); pte 284 arch/x86/mm/pti.c if (!pte || WARN_ON(level != PG_LEVEL_4K) || pte_none(*pte)) pte 291 arch/x86/mm/pti.c *target_pte = *pte; pte 314 arch/x86/mm/pti.c pte_t *pte, *target_pte; pte 383 arch/x86/mm/pti.c pte = pte_offset_kernel(pmd, addr); pte 384 arch/x86/mm/pti.c if (pte_none(*pte)) { pte 390 arch/x86/mm/pti.c if (WARN_ON(!(pte_flags(*pte) & _PAGE_PRESENT))) pte 400 arch/x86/mm/pti.c *pte = pte_set_flags(*pte, _PAGE_GLOBAL); pte 403 arch/x86/mm/pti.c *target_pte = *pte; pte 217 arch/x86/power/hibernate.c pte_t *pte; pte 243 arch/x86/power/hibernate.c pte = pte_offset_kernel(pmd, relocated_restore_code); pte 244 arch/x86/power/hibernate.c set_pte(pte, __pte(pte_val(*pte) & ~_PAGE_NX)); pte 86 arch/x86/power/hibernate_32.c pte_t *pte; pte 115 arch/x86/power/hibernate_32.c pte = resume_one_page_table_init(pmd); pte 116 arch/x86/power/hibernate_32.c if (!pte) pte 119 arch/x86/power/hibernate_32.c max_pte = pte + PTRS_PER_PTE; pte 120 arch/x86/power/hibernate_32.c for (; pte < max_pte; pte++, pfn++) { pte 124 arch/x86/power/hibernate_32.c set_pte(pte, pfn_pte(pfn, PAGE_KERNEL_EXEC)); pte 149 arch/x86/power/hibernate_32.c pte_t *pte; pte 161 arch/x86/power/hibernate_32.c pte = resume_one_page_table_init(pmd); pte 162 arch/x86/power/hibernate_32.c if (!pte) pte 164 arch/x86/power/hibernate_32.c set_pte(pte + pte_index(restore_jump_address), pte 352 arch/x86/xen/enlighten_pv.c pte_t pte; pte 363 arch/x86/xen/enlighten_pv.c pte = pfn_pte(pfn, prot); pte 389 arch/x86/xen/enlighten_pv.c if (HYPERVISOR_update_va_mapping((unsigned long)v, pte, 0)) pte 396 arch/x86/xen/enlighten_pv.c if (HYPERVISOR_update_va_mapping((unsigned long)av, pte, 0)) pte 492 arch/x86/xen/enlighten_pv.c pte_t pte; pte 501 arch/x86/xen/enlighten_pv.c pte = pfn_pte(pfn, PAGE_KERNEL_RO); pte 503 arch/x86/xen/enlighten_pv.c if (HYPERVISOR_update_va_mapping((unsigned long)va, pte, 0)) pte 22 arch/x86/xen/mmu.c pte_t *pte; pte 34 arch/x86/xen/mmu.c pte = lookup_address(address, &level); pte 35 arch/x86/xen/mmu.c BUG_ON(pte == NULL); pte 37 arch/x86/xen/mmu.c return XMADDR(((phys_addr_t)pte_mfn(*pte) << PAGE_SHIFT) + offset); pte 22 arch/x86/xen/mmu.h pte_t *ptep, pte_t pte); pte 138 arch/x86/xen/mmu_pv.c pte_t *pte, ptev; pte 142 arch/x86/xen/mmu_pv.c pte = lookup_address(address, &level); pte 143 arch/x86/xen/mmu_pv.c if (pte == NULL) pte 146 arch/x86/xen/mmu_pv.c ptev = pte_wrprotect(*pte); pte 154 arch/x86/xen/mmu_pv.c pte_t *pte, ptev; pte 158 arch/x86/xen/mmu_pv.c pte = lookup_address(address, &level); pte 159 arch/x86/xen/mmu_pv.c if (pte == NULL) pte 162 arch/x86/xen/mmu_pv.c ptev = pte_mkwrite(*pte); pte 318 arch/x86/xen/mmu_pv.c pte_t *ptep, pte_t pte) pte 322 arch/x86/xen/mmu_pv.c trace_xen_mmu_ptep_modify_prot_commit(vma->vm_mm, addr, ptep, pte); pte 326 arch/x86/xen/mmu_pv.c u.val = pte_val_ma(pte); pte 375 arch/x86/xen/mmu_pv.c __visible pteval_t xen_pte_val(pte_t pte) pte 377 arch/x86/xen/mmu_pv.c pteval_t pteval = pte.pte; pte 389 arch/x86/xen/mmu_pv.c __visible pte_t xen_make_pte(pteval_t pte) pte 391 arch/x86/xen/mmu_pv.c pte = pte_pfn_to_mfn(pte); pte 393 arch/x86/xen/mmu_pv.c return native_make_pte(pte); pte 443 arch/x86/xen/mmu_pv.c static void xen_set_pte_atomic(pte_t *ptep, pte_t pte) pte 445 arch/x86/xen/mmu_pv.c trace_xen_mmu_set_pte_atomic(ptep, pte); pte 446 arch/x86/xen/mmu_pv.c __xen_set_pte(ptep, pte); pte 1540 arch/x86/xen/mmu_pv.c __visible pte_t xen_make_pte_init(pteval_t pte) pte 1551 arch/x86/xen/mmu_pv.c pfn = (pte & PTE_PFN_MASK) >> PAGE_SHIFT; pte 1555 arch/x86/xen/mmu_pv.c pte &= ~_PAGE_RW; pte 1557 arch/x86/xen/mmu_pv.c pte = pte_pfn_to_mfn(pte); pte 1558 arch/x86/xen/mmu_pv.c return native_make_pte(pte); pte 1562 arch/x86/xen/mmu_pv.c static void __init xen_set_pte_init(pte_t *ptep, pte_t pte) pte 1566 arch/x86/xen/mmu_pv.c if (pte_mfn(pte) != INVALID_P2M_ENTRY pte 1568 arch/x86/xen/mmu_pv.c pte = __pte_ma(((pte_val_ma(*ptep) & _PAGE_RW) | ~_PAGE_RW) & pte 1569 arch/x86/xen/mmu_pv.c pte_val_ma(pte)); pte 1571 arch/x86/xen/mmu_pv.c __xen_set_pte(ptep, pte); pte 1765 arch/x86/xen/mmu_pv.c pte_t pte = pfn_pte(pfn, prot); pte 1767 arch/x86/xen/mmu_pv.c if (HYPERVISOR_update_va_mapping((unsigned long)addr, pte, flags)) pte 1805 arch/x86/xen/mmu_pv.c pte_t pte; pte 1813 arch/x86/xen/mmu_pv.c pte = pfn_pte(pfn, PAGE_KERNEL_EXEC); pte 1814 arch/x86/xen/mmu_pv.c pte_page[pteidx] = pte; pte 1843 arch/x86/xen/mmu_pv.c pte_t *pte = v; pte 1849 arch/x86/xen/mmu_pv.c pte[i] = xen_make_pte(pte[i].pte); pte 2018 arch/x86/xen/mmu_pv.c pte_t pte; pte 2043 arch/x86/xen/mmu_pv.c pte = native_make_pte(xen_read_phys_ulong(pa + pte_index(vaddr) * pte 2044 arch/x86/xen/mmu_pv.c sizeof(pte))); pte 2045 arch/x86/xen/mmu_pv.c if (!pte_present(pte)) pte 2047 arch/x86/xen/mmu_pv.c pa = pte_pfn(pte) << PAGE_SHIFT; pte 2297 arch/x86/xen/mmu_pv.c pte_t pte; pte 2312 arch/x86/xen/mmu_pv.c pte = pfn_pte(phys, prot); pte 2317 arch/x86/xen/mmu_pv.c pte = pfn_pte(PFN_DOWN(__pa(dummy_mapping)), PAGE_KERNEL); pte 2327 arch/x86/xen/mmu_pv.c pte = pfn_pte(PFN_DOWN(__pa(dummy_mapping)), PAGE_KERNEL); pte 2334 arch/x86/xen/mmu_pv.c pte = mfn_pte(phys, prot); pte 2339 arch/x86/xen/mmu_pv.c pte = mfn_pte(phys, prot); pte 2343 arch/x86/xen/mmu_pv.c __native_set_fixmap(idx, pte); pte 2350 arch/x86/xen/mmu_pv.c set_pte_vaddr_pud(level3_user_vsyscall, vaddr, pte); pte 2694 arch/x86/xen/mmu_pv.c pte_t pte = pte_mkspecial(mfn_pte(*rmd->pfn, rmd->prot)); pte 2709 arch/x86/xen/mmu_pv.c rmd->mmu_update->val = pte_val_ma(pte); pte 703 arch/x86/xen/p2m.c pte_t *pte; pte 723 arch/x86/xen/p2m.c pte = (pte_t *)(mfn_to_virt(PFN_DOWN(map_ops[i].host_addr)) + pte 725 arch/x86/xen/p2m.c mfn = pte_mfn(*pte); pte 97 arch/xtensa/include/asm/page.h typedef struct { unsigned long pte; } pte_t; /* page table entry */ pte 102 arch/xtensa/include/asm/page.h #define pte_val(x) ((x).pte) pte 51 arch/xtensa/include/asm/pgalloc.h pte_t *pte; pte 54 arch/xtensa/include/asm/pgalloc.h pte = pte_alloc_one_kernel(mm); pte 55 arch/xtensa/include/asm/pgalloc.h if (!pte) pte 57 arch/xtensa/include/asm/pgalloc.h page = virt_to_page(pte); pte 65 arch/xtensa/include/asm/pgalloc.h static inline void pte_free_kernel(struct mm_struct *mm, pte_t *pte) pte 67 arch/xtensa/include/asm/pgalloc.h free_page((unsigned long)pte); pte 70 arch/xtensa/include/asm/pgalloc.h static inline void pte_free(struct mm_struct *mm, pgtable_t pte) pte 72 arch/xtensa/include/asm/pgalloc.h pgtable_pte_page_dtor(pte); pte 73 arch/xtensa/include/asm/pgalloc.h __free_page(pte); pte 251 arch/xtensa/include/asm/pgtable.h # define pte_none(pte) (pte_val(pte) == (_PAGE_CA_INVALID | _PAGE_USER)) pte 253 arch/xtensa/include/asm/pgtable.h # define pte_present(pte) ((pte_val(pte) & _PAGE_CA_MASK) != _PAGE_CA_INVALID) pte 255 arch/xtensa/include/asm/pgtable.h # define pte_present(pte) \ pte 256 arch/xtensa/include/asm/pgtable.h (((pte_val(pte) & _PAGE_CA_MASK) != _PAGE_CA_INVALID) \ pte 257 arch/xtensa/include/asm/pgtable.h || ((pte_val(pte) & _PAGE_ATTRIB_MASK) == _PAGE_NONE)) pte 267 arch/xtensa/include/asm/pgtable.h static inline int pte_write(pte_t pte) { return pte_val(pte) & _PAGE_WRITABLE; } pte 268 arch/xtensa/include/asm/pgtable.h static inline int pte_dirty(pte_t pte) { return pte_val(pte) & _PAGE_DIRTY; } pte 269 arch/xtensa/include/asm/pgtable.h static inline int pte_young(pte_t pte) { return pte_val(pte) & _PAGE_ACCESSED; } pte 270 arch/xtensa/include/asm/pgtable.h static inline int pte_special(pte_t pte) { return 0; } pte 272 arch/xtensa/include/asm/pgtable.h static inline pte_t pte_wrprotect(pte_t pte) pte 273 arch/xtensa/include/asm/pgtable.h { pte_val(pte) &= ~(_PAGE_WRITABLE | _PAGE_HW_WRITE); return pte; } pte 274 arch/xtensa/include/asm/pgtable.h static inline pte_t pte_mkclean(pte_t pte) pte 275 arch/xtensa/include/asm/pgtable.h { pte_val(pte) &= ~(_PAGE_DIRTY | _PAGE_HW_WRITE); return pte; } pte 276 arch/xtensa/include/asm/pgtable.h static inline pte_t pte_mkold(pte_t pte) pte 277 arch/xtensa/include/asm/pgtable.h { pte_val(pte) &= ~_PAGE_ACCESSED; return pte; } pte 278 arch/xtensa/include/asm/pgtable.h static inline pte_t pte_mkdirty(pte_t pte) pte 279 arch/xtensa/include/asm/pgtable.h { pte_val(pte) |= _PAGE_DIRTY; return pte; } pte 280 arch/xtensa/include/asm/pgtable.h static inline pte_t pte_mkyoung(pte_t pte) pte 281 arch/xtensa/include/asm/pgtable.h { pte_val(pte) |= _PAGE_ACCESSED; return pte; } pte 282 arch/xtensa/include/asm/pgtable.h static inline pte_t pte_mkwrite(pte_t pte) pte 283 arch/xtensa/include/asm/pgtable.h { pte_val(pte) |= _PAGE_WRITABLE; return pte; } pte 284 arch/xtensa/include/asm/pgtable.h static inline pte_t pte_mkspecial(pte_t pte) pte 285 arch/xtensa/include/asm/pgtable.h { return pte; } pte 294 arch/xtensa/include/asm/pgtable.h #define pte_pfn(pte) (pte_val(pte) >> PAGE_SHIFT) pte 300 arch/xtensa/include/asm/pgtable.h static inline pte_t pte_modify(pte_t pte, pgprot_t newprot) pte 302 arch/xtensa/include/asm/pgtable.h return __pte((pte_val(pte) & _PAGE_CHG_MASK) | pgprot_val(newprot)); pte 344 arch/xtensa/include/asm/pgtable.h pte_t pte = *ptep; pte 345 arch/xtensa/include/asm/pgtable.h if (!pte_young(pte)) pte 347 arch/xtensa/include/asm/pgtable.h update_pte(ptep, pte_mkold(pte)); pte 354 arch/xtensa/include/asm/pgtable.h pte_t pte = *ptep; pte 356 arch/xtensa/include/asm/pgtable.h return pte; pte 362 arch/xtensa/include/asm/pgtable.h pte_t pte = *ptep; pte 363 arch/xtensa/include/asm/pgtable.h update_pte(ptep, pte_wrprotect(pte)); pte 382 arch/xtensa/include/asm/pgtable.h #define pte_unmap(pte) do { } while (0) pte 396 arch/xtensa/include/asm/pgtable.h #define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val(pte) }) pte 19 arch/xtensa/include/asm/tlb.h #define __pte_free_tlb(tlb, pte, address) pte_free((tlb)->mm, pte) pte 46 arch/xtensa/mm/kasan_init.c pte_t *pte = memblock_alloc(n_pages * sizeof(pte_t), PAGE_SIZE); pte 48 arch/xtensa/mm/kasan_init.c if (!pte) pte 66 arch/xtensa/mm/kasan_init.c set_pte(pte + j, pfn_pte(PHYS_PFN(phys), PAGE_KERNEL)); pte 70 arch/xtensa/mm/kasan_init.c for (i = 0; i < n_pmds ; ++i, pte += PTRS_PER_PTE) pte 71 arch/xtensa/mm/kasan_init.c set_pmd(pmd + i, __pmd((unsigned long)pte)); pte 26 arch/xtensa/mm/mmu.c pte_t *pte; pte 34 arch/xtensa/mm/mmu.c pte = memblock_alloc_low(n_pages * sizeof(pte_t), PAGE_SIZE); pte 35 arch/xtensa/mm/mmu.c if (!pte) pte 40 arch/xtensa/mm/mmu.c pte_clear(NULL, 0, pte + i); pte 43 arch/xtensa/mm/mmu.c pte_t *cur_pte = pte + i; pte 51 arch/xtensa/mm/mmu.c return pte; pte 173 arch/xtensa/mm/tlb.c pte_t *pte; pte 183 arch/xtensa/mm/tlb.c pte = pte_offset_map(pmd, vaddr); pte 184 arch/xtensa/mm/tlb.c if (!pte) pte 186 arch/xtensa/mm/tlb.c return pte_val(*pte); pte 222 arch/xtensa/mm/tlb.c unsigned pte = get_pte_for_vaddr(vpn); pte 236 arch/xtensa/mm/tlb.c if ((pte ^ r1) & PAGE_MASK) { pte 238 arch/xtensa/mm/tlb.c dtlb ? 'D' : 'I', w, e, r0, r1, pte); pte 239 arch/xtensa/mm/tlb.c if (pte == 0 || !pte_present(__pte(pte))) { pte 277 block/partitions/efi.c gpt_entry *pte; pte 286 block/partitions/efi.c pte = kmalloc(count, GFP_KERNEL); pte 287 block/partitions/efi.c if (!pte) pte 291 block/partitions/efi.c (u8 *) pte, count) < count) { pte 292 block/partitions/efi.c kfree(pte); pte 293 block/partitions/efi.c pte=NULL; pte 296 block/partitions/efi.c return pte; pte 463 block/partitions/efi.c is_pte_valid(const gpt_entry *pte, const u64 lastlba) pte 465 block/partitions/efi.c if ((!efi_guidcmp(pte->partition_type_guid, NULL_GUID)) || pte 466 block/partitions/efi.c le64_to_cpu(pte->starting_lba) > lastlba || pte 467 block/partitions/efi.c le64_to_cpu(pte->ending_lba) > lastlba) pte 51 drivers/char/agp/amd64-agp.c u32 pte; pte 87 drivers/char/agp/amd64-agp.c pte = (tmp & 0x000000ff00000000ULL) >> 28; pte 88 drivers/char/agp/amd64-agp.c pte |=(tmp & 0x00000000fffff000ULL); pte 89 drivers/char/agp/amd64-agp.c pte |= GPTE_VALID | GPTE_COHERENT; pte 91 drivers/char/agp/amd64-agp.c writel(pte, agp_bridge->gatt_table+j); pte 201 drivers/gpu/drm/amd/amdgpu/amdgpu_vm_sdma.c uint64_t *pte; pte 248 drivers/gpu/drm/amd/amdgpu/amdgpu_vm_sdma.c pte = (uint64_t *)&(p->job->ibs->ptr[p->num_dw_left]); pte 250 drivers/gpu/drm/amd/amdgpu/amdgpu_vm_sdma.c pte[i] = amdgpu_vm_map_gart(p->pages_addr, addr); pte 251 drivers/gpu/drm/amd/amdgpu/amdgpu_vm_sdma.c pte[i] |= flags; pte 142 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.c const struct pte_setting *pte = &pte_settings[mi_tiling][mi_bpp]; pte 144 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.c unsigned int page_width = log_2(pte->page_width); pte 145 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.c unsigned int page_height = log_2(pte->page_height); pte 147 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.c pte->min_pte_before_flip_vert_scan : pte 148 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.c pte->min_pte_before_flip_horiz_scan; pte 159 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.c DVMM_PTE_REQ_PER_CHUNK, pte->pte_req_per_chunk, pte 570 drivers/gpu/drm/amd/display/dc/dce110/dce110_mem_input_v.c const unsigned int *pte = get_dvmm_hw_setting(tiling_info, format, false); pte 577 drivers/gpu/drm/amd/display/dc/dce110/dce110_mem_input_v.c unsigned int temp_page_width = pte[1]; pte 578 drivers/gpu/drm/amd/display/dc/dce110/dce110_mem_input_v.c unsigned int temp_page_height = pte[2]; pte 598 drivers/gpu/drm/amd/display/dc/dce110/dce110_mem_input_v.c min_pte_before_flip = pte[4]; pte 602 drivers/gpu/drm/amd/display/dc/dce110/dce110_mem_input_v.c min_pte_before_flip = pte[3]; pte 620 drivers/gpu/drm/amd/display/dc/dce110/dce110_mem_input_v.c set_reg_field_value(value, pte[5], UNP_DVMM_PTE_ARB_CONTROL, DVMM_PTE_REQ_PER_CHUNK); pte 79 drivers/gpu/drm/gma500/gtt.c u32 pte; pte 100 drivers/gpu/drm/gma500/gtt.c pte = psb_gtt_mask_pte(page_to_pfn(r->pages[i]), pte 102 drivers/gpu/drm/gma500/gtt.c iowrite32(pte, gtt_slot++); pte 105 drivers/gpu/drm/gma500/gtt.c pte = psb_gtt_mask_pte(page_to_pfn(r->pages[i]), pte 107 drivers/gpu/drm/gma500/gtt.c iowrite32(pte, gtt_slot++); pte 128 drivers/gpu/drm/gma500/gtt.c u32 pte; pte 134 drivers/gpu/drm/gma500/gtt.c pte = psb_gtt_mask_pte(page_to_pfn(dev_priv->scratch_page), pte 138 drivers/gpu/drm/gma500/gtt.c iowrite32(pte, gtt_slot++); pte 156 drivers/gpu/drm/gma500/gtt.c u32 pte; pte 174 drivers/gpu/drm/gma500/gtt.c pte = psb_gtt_mask_pte(page_to_pfn(r->pages[i]), pte 176 drivers/gpu/drm/gma500/gtt.c iowrite32(pte, gtt_slot++); pte 179 drivers/gpu/drm/gma500/gtt.c pte = psb_gtt_mask_pte(page_to_pfn(r->pages[i]), pte 181 drivers/gpu/drm/gma500/gtt.c iowrite32(pte, gtt_slot++); pte 416 drivers/gpu/drm/gma500/gtt.c uint32_t pte; pte 533 drivers/gpu/drm/gma500/gtt.c pte = psb_gtt_mask_pte(pfn_base + i, PSB_MMU_CACHED_MEMORY); pte 534 drivers/gpu/drm/gma500/gtt.c iowrite32(pte, dev_priv->gtt_map + i); pte 542 drivers/gpu/drm/gma500/gtt.c pte = psb_gtt_mask_pte(pfn_base, PSB_MMU_CACHED_MEMORY); pte 544 drivers/gpu/drm/gma500/gtt.c iowrite32(pte, dev_priv->gtt_map + i); pte 397 drivers/gpu/drm/gma500/mmu.c uint32_t pte) pte 399 drivers/gpu/drm/gma500/mmu.c pt->v[psb_mmu_pt_index(addr)] = pte; pte 655 drivers/gpu/drm/gma500/mmu.c uint32_t pte; pte 675 drivers/gpu/drm/gma500/mmu.c pte = psb_mmu_mask_pte(start_pfn++, type); pte 676 drivers/gpu/drm/gma500/mmu.c psb_mmu_set_pte(pt, addr, pte); pte 704 drivers/gpu/drm/gma500/mmu.c uint32_t pte; pte 737 drivers/gpu/drm/gma500/mmu.c pte = psb_mmu_mask_pte(page_to_pfn(*pages++), pte 739 drivers/gpu/drm/gma500/mmu.c psb_mmu_set_pte(pt, addr, pte); pte 37 drivers/gpu/drm/i915/gvt/dmabuf.c #define GEN8_DECODE_PTE(pte) (pte & GENMASK_ULL(63, 12)) pte 293 drivers/gpu/drm/i915/gvt/gtt.c unsigned long index, u64 pte) pte 297 drivers/gpu/drm/i915/gvt/gtt.c writeq(pte, addr); pte 493 drivers/gpu/drm/i915/gvt/gtt.c DEFINE_PPGTT_GMA_TO_INDEX(gen8, pte, (gma >> 12 & 0x1ff)); pte 202 drivers/gpu/drm/i915/i915_gem_gtt.c gen8_pte_t pte = addr | _PAGE_PRESENT | _PAGE_RW; pte 205 drivers/gpu/drm/i915/i915_gem_gtt.c pte &= ~_PAGE_RW; pte 209 drivers/gpu/drm/i915/i915_gem_gtt.c pte |= PPAT_UNCACHED; pte 212 drivers/gpu/drm/i915/i915_gem_gtt.c pte |= PPAT_DISPLAY_ELLC; pte 215 drivers/gpu/drm/i915/i915_gem_gtt.c pte |= PPAT_CACHED; pte 219 drivers/gpu/drm/i915/i915_gem_gtt.c return pte; pte 238 drivers/gpu/drm/i915/i915_gem_gtt.c gen6_pte_t pte = GEN6_PTE_VALID; pte 239 drivers/gpu/drm/i915/i915_gem_gtt.c pte |= GEN6_PTE_ADDR_ENCODE(addr); pte 244 drivers/gpu/drm/i915/i915_gem_gtt.c pte |= GEN6_PTE_CACHE_LLC; pte 247 drivers/gpu/drm/i915/i915_gem_gtt.c pte |= GEN6_PTE_UNCACHED; pte 253 drivers/gpu/drm/i915/i915_gem_gtt.c return pte; pte 260 drivers/gpu/drm/i915/i915_gem_gtt.c gen6_pte_t pte = GEN6_PTE_VALID; pte 261 drivers/gpu/drm/i915/i915_gem_gtt.c pte |= GEN6_PTE_ADDR_ENCODE(addr); pte 265 drivers/gpu/drm/i915/i915_gem_gtt.c pte |= GEN7_PTE_CACHE_L3_LLC; pte 268 drivers/gpu/drm/i915/i915_gem_gtt.c pte |= GEN6_PTE_CACHE_LLC; pte 271 drivers/gpu/drm/i915/i915_gem_gtt.c pte |= GEN6_PTE_UNCACHED; pte 277 drivers/gpu/drm/i915/i915_gem_gtt.c return pte; pte 284 drivers/gpu/drm/i915/i915_gem_gtt.c gen6_pte_t pte = GEN6_PTE_VALID; pte 285 drivers/gpu/drm/i915/i915_gem_gtt.c pte |= GEN6_PTE_ADDR_ENCODE(addr); pte 288 drivers/gpu/drm/i915/i915_gem_gtt.c pte |= BYT_PTE_WRITEABLE; pte 291 drivers/gpu/drm/i915/i915_gem_gtt.c pte |= BYT_PTE_SNOOPED_BY_CPU_CACHES; pte 293 drivers/gpu/drm/i915/i915_gem_gtt.c return pte; pte 300 drivers/gpu/drm/i915/i915_gem_gtt.c gen6_pte_t pte = GEN6_PTE_VALID; pte 301 drivers/gpu/drm/i915/i915_gem_gtt.c pte |= HSW_PTE_ADDR_ENCODE(addr); pte 304 drivers/gpu/drm/i915/i915_gem_gtt.c pte |= HSW_WB_LLC_AGE3; pte 306 drivers/gpu/drm/i915/i915_gem_gtt.c return pte; pte 313 drivers/gpu/drm/i915/i915_gem_gtt.c gen6_pte_t pte = GEN6_PTE_VALID; pte 314 drivers/gpu/drm/i915/i915_gem_gtt.c pte |= HSW_PTE_ADDR_ENCODE(addr); pte 320 drivers/gpu/drm/i915/i915_gem_gtt.c pte |= HSW_WT_ELLC_LLC_AGE3; pte 323 drivers/gpu/drm/i915/i915_gem_gtt.c pte |= HSW_WB_ELLC_LLC_AGE3; pte 327 drivers/gpu/drm/i915/i915_gem_gtt.c return pte; pte 1611 drivers/gpu/drm/i915/i915_gem_gtt.c unsigned int pte = first_entry % GEN6_PTES; pte 1617 drivers/gpu/drm/i915/i915_gem_gtt.c const unsigned int count = min(num_entries, GEN6_PTES - pte); pte 1636 drivers/gpu/drm/i915/i915_gem_gtt.c memset32(vaddr + pte, scratch_pte, count); pte 1639 drivers/gpu/drm/i915/i915_gem_gtt.c pte = 0; pte 2169 drivers/gpu/drm/i915/i915_gem_gtt.c static void gen8_set_pte(void __iomem *addr, gen8_pte_t pte) pte 2171 drivers/gpu/drm/i915/i915_gem_gtt.c writeq(pte, addr); pte 2181 drivers/gpu/drm/i915/i915_gem_gtt.c gen8_pte_t __iomem *pte = pte 2184 drivers/gpu/drm/i915/i915_gem_gtt.c gen8_set_pte(pte, gen8_pte_encode(addr, level, 0)); pte 2224 drivers/gpu/drm/i915/i915_gem_gtt.c gen6_pte_t __iomem *pte = pte 2227 drivers/gpu/drm/i915/i915_gem_gtt.c iowrite32(vm->pte_encode(addr, level, flags), pte); pte 38 drivers/gpu/drm/i915/i915_mm.c static int remap_pfn(pte_t *pte, unsigned long addr, void *data) pte 43 drivers/gpu/drm/i915/i915_mm.c set_pte_at(r->mm, addr, pte, pte_mkspecial(pfn_pte(r->pfn, r->prot))); pte 275 drivers/gpu/drm/lima/lima_vm.c u32 pte = *pt++; pte 277 drivers/gpu/drm/lima/lima_vm.c if (pte) pte 278 drivers/gpu/drm/lima/lima_vm.c printk(KERN_INFO " pt %03x:%08x\n", k, pte); pte 78 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c u32 pte[NVKM_VMM_LEVELS_MAX]; pte 103 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c buf += sprintf(buf, "%05x:", it->pte[lvl]); pte 146 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c u32 pdei = it->pte[it->lvl + 1]; pte 211 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c pgt->pte[lpti] -= pten; pte 221 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c if (pgt->pte[pteb] & NVKM_VMM_PTE_SPTES) { pte 223 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c if (!(pgt->pte[ptei] & NVKM_VMM_PTE_SPTES)) pte 235 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c pgt->pte[ptei] &= ~NVKM_VMM_PTE_VALID; pte 237 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c if (pgt->pte[ptei] & NVKM_VMM_PTE_SPTES) pte 239 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c pgt->pte[ptei] &= ~NVKM_VMM_PTE_VALID; pte 242 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c if (pgt->pte[pteb] & NVKM_VMM_PTE_SPARSE) { pte 310 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c pgt->pte[lpti] += pten; pte 320 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c if (pgt->pte[pteb] & NVKM_VMM_PTE_VALID) { pte 322 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c if (!(pgt->pte[ptei] & NVKM_VMM_PTE_VALID)) pte 334 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c pgt->pte[ptei] |= NVKM_VMM_PTE_VALID; pte 336 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c if (pgt->pte[ptei] & NVKM_VMM_PTE_VALID) pte 338 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c pgt->pte[ptei] |= NVKM_VMM_PTE_VALID; pte 341 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c if (pgt->pte[pteb] & NVKM_VMM_PTE_SPARSE) { pte 389 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c memset(&pgt->pte[ptei], NVKM_VMM_PTE_SPARSE, ptes); pte 401 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c memset(&pt->pte[ptei], 0x00, sizeof(pt->pte[0]) * ptes); pte 448 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c bool spte = pgt->pte[ptei] & NVKM_VMM_PTE_SPTES; pte 450 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c bool next = pgt->pte[ptei] & NVKM_VMM_PTE_SPTES; pte 460 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c memset(&pgt->pte[pteb], 0x00, ptes); pte 464 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c pgt->pte[pteb++] |= NVKM_VMM_PTE_VALID; pte 519 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c it.pte[it.lvl] = bits & ((1 << desc[it.lvl].bits) - 1); pte 535 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c const u32 ptei = it.pte[0]; pte 540 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c const u32 pdei = it.pte[it.lvl]; pte 575 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c it.pte[it.lvl] += ptes; pte 578 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c while (it.pte[it.lvl] == (1 << desc[it.lvl].bits)) { pte 579 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c it.pte[it.lvl++] = 0; pte 580 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c it.pte[it.lvl]++; pte 592 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c addr = it.pte[it.max--]; pte 595 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c addr |= it.pte[it.max]; pte 50 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.h u8 pte[]; pte 830 drivers/gpu/drm/radeon/radeon_vm.c uint64_t pte; pte 843 drivers/gpu/drm/radeon/radeon_vm.c pte = radeon_bo_gpu_offset(pt); pte 844 drivers/gpu/drm/radeon/radeon_vm.c pte += (addr & mask) * 8; pte 846 drivers/gpu/drm/radeon/radeon_vm.c if ((last_pte + 8 * count) != pte) { pte 855 drivers/gpu/drm/radeon/radeon_vm.c last_pte = pte; pte 98 drivers/gpu/drm/v3d/v3d_mmu.c u32 pte = page_prot | page_address; pte 105 drivers/gpu/drm/v3d/v3d_mmu.c v3d->pt[page++] = pte + i; pte 490 drivers/iommu/amd_iommu.c static u64 *first_pte_l7(u64 *pte, unsigned long *page_size, pte 496 drivers/iommu/amd_iommu.c pg_size = PTE_PAGE_SIZE(*pte); pte 499 drivers/iommu/amd_iommu.c fpte = (u64 *)(((unsigned long)pte) & pte_mask); pte 1470 drivers/iommu/amd_iommu.c u64 *pte; pte 1478 drivers/iommu/amd_iommu.c pte = (void *)get_zeroed_page(gfp); pte 1479 drivers/iommu/amd_iommu.c if (!pte) pte 1482 drivers/iommu/amd_iommu.c *pte = PM_LEVEL_PDE(domain->mode, pte 1484 drivers/iommu/amd_iommu.c domain->pt_root = pte; pte 1503 drivers/iommu/amd_iommu.c u64 *pte, *page; pte 1511 drivers/iommu/amd_iommu.c pte = &domain->pt_root[PM_LEVEL_INDEX(level, address)]; pte 1519 drivers/iommu/amd_iommu.c __pte = *pte; pte 1531 drivers/iommu/amd_iommu.c lpte = first_pte_l7(pte, NULL, &count); pte 1554 drivers/iommu/amd_iommu.c if (cmpxchg64(pte, __pte, __npte) != __pte) pte 1568 drivers/iommu/amd_iommu.c pte = IOMMU_PTE_PAGE(__pte); pte 1571 drivers/iommu/amd_iommu.c *pte_page = pte; pte 1573 drivers/iommu/amd_iommu.c pte = &pte[PM_LEVEL_INDEX(level, address)]; pte 1576 drivers/iommu/amd_iommu.c return pte; pte 1588 drivers/iommu/amd_iommu.c u64 *pte; pte 1596 drivers/iommu/amd_iommu.c pte = &domain->pt_root[PM_LEVEL_INDEX(level, address)]; pte 1602 drivers/iommu/amd_iommu.c if (!IOMMU_PTE_PRESENT(*pte)) pte 1606 drivers/iommu/amd_iommu.c if (PM_PTE_LEVEL(*pte) == 7 || pte 1607 drivers/iommu/amd_iommu.c PM_PTE_LEVEL(*pte) == 0) pte 1611 drivers/iommu/amd_iommu.c if (PM_PTE_LEVEL(*pte) != level) pte 1617 drivers/iommu/amd_iommu.c pte = IOMMU_PTE_PAGE(*pte); pte 1618 drivers/iommu/amd_iommu.c pte = &pte[PM_LEVEL_INDEX(level, address)]; pte 1626 drivers/iommu/amd_iommu.c if (PM_PTE_LEVEL(*pte) == PAGE_MODE_7_LEVEL) pte 1627 drivers/iommu/amd_iommu.c pte = first_pte_l7(pte, page_size, NULL); pte 1629 drivers/iommu/amd_iommu.c return pte; pte 1632 drivers/iommu/amd_iommu.c static struct page *free_clear_pte(u64 *pte, u64 pteval, struct page *freelist) pte 1637 drivers/iommu/amd_iommu.c while (cmpxchg64(pte, pteval, 0) != pteval) { pte 1639 drivers/iommu/amd_iommu.c pteval = *pte; pte 1667 drivers/iommu/amd_iommu.c u64 __pte, *pte; pte 1678 drivers/iommu/amd_iommu.c pte = alloc_pte(dom, bus_addr, page_size, NULL, gfp, &updated); pte 1681 drivers/iommu/amd_iommu.c if (!pte) pte 1685 drivers/iommu/amd_iommu.c freelist = free_clear_pte(&pte[i], pte[i], freelist); pte 1702 drivers/iommu/amd_iommu.c pte[i] = __pte; pte 1727 drivers/iommu/amd_iommu.c u64 *pte; pte 1735 drivers/iommu/amd_iommu.c pte = fetch_pte(dom, bus_addr, &unmap_size); pte 1737 drivers/iommu/amd_iommu.c if (pte) { pte 1742 drivers/iommu/amd_iommu.c pte[i] = 0ULL; pte 3144 drivers/iommu/amd_iommu.c u64 *pte, __pte; pte 3149 drivers/iommu/amd_iommu.c pte = fetch_pte(domain, iova, &pte_pgsize); pte 3151 drivers/iommu/amd_iommu.c if (!pte || !IOMMU_PTE_PRESENT(*pte)) pte 3155 drivers/iommu/amd_iommu.c __pte = __sme_clr(*pte & PM_ADDR_MASK); pte 3489 drivers/iommu/amd_iommu.c u64 *pte; pte 3494 drivers/iommu/amd_iommu.c pte = &root[index]; pte 3499 drivers/iommu/amd_iommu.c if (!(*pte & GCR3_VALID)) { pte 3507 drivers/iommu/amd_iommu.c *pte = iommu_virt_to_phys(root) | GCR3_VALID; pte 3510 drivers/iommu/amd_iommu.c root = iommu_phys_to_virt(*pte & PAGE_MASK); pte 3515 drivers/iommu/amd_iommu.c return pte; pte 3521 drivers/iommu/amd_iommu.c u64 *pte; pte 3526 drivers/iommu/amd_iommu.c pte = __get_gcr3_pte(domain->gcr3_tbl, domain->glx, pasid, true); pte 3527 drivers/iommu/amd_iommu.c if (pte == NULL) pte 3530 drivers/iommu/amd_iommu.c *pte = (cr3 & PAGE_MASK) | GCR3_VALID; pte 3537 drivers/iommu/amd_iommu.c u64 *pte; pte 3542 drivers/iommu/amd_iommu.c pte = __get_gcr3_pte(domain->gcr3_tbl, domain->glx, pasid, false); pte 3543 drivers/iommu/amd_iommu.c if (pte == NULL) pte 3546 drivers/iommu/amd_iommu.c *pte = 0; pte 277 drivers/iommu/amd_iommu_types.h #define PM_PTE_LEVEL(pte) (((pte) >> 9) & 0x7ULL) pte 316 drivers/iommu/amd_iommu_types.h #define PTE_PAGE_SIZE(pte) \ pte 317 drivers/iommu/amd_iommu_types.h (1ULL << (1 + ffz(((pte) | 0xfffULL)))) pte 364 drivers/iommu/amd_iommu_types.h #define IOMMU_PTE_PRESENT(pte) ((pte) & IOMMU_PTE_PR) pte 365 drivers/iommu/amd_iommu_types.h #define IOMMU_PTE_PAGE(pte) (iommu_phys_to_virt((pte) & IOMMU_PAGE_MASK)) pte 366 drivers/iommu/amd_iommu_types.h #define IOMMU_PTE_MODE(pte) (((pte) >> 9) & 0x07) pte 884 drivers/iommu/intel-iommu.c struct dma_pte *parent, *pte; pte 900 drivers/iommu/intel-iommu.c pte = &parent[offset]; pte 901 drivers/iommu/intel-iommu.c if (!*target_level && (dma_pte_superpage(pte) || !dma_pte_present(pte))) pte 906 drivers/iommu/intel-iommu.c if (!dma_pte_present(pte)) { pte 916 drivers/iommu/intel-iommu.c if (cmpxchg64(&pte->val, 0ULL, pteval)) pte 920 drivers/iommu/intel-iommu.c domain_flush_cache(domain, pte, sizeof(*pte)); pte 925 drivers/iommu/intel-iommu.c parent = phys_to_virt(dma_pte_addr(pte)); pte 932 drivers/iommu/intel-iommu.c return pte; pte 940 drivers/iommu/intel-iommu.c struct dma_pte *parent, *pte; pte 947 drivers/iommu/intel-iommu.c pte = &parent[offset]; pte 949 drivers/iommu/intel-iommu.c return pte; pte 951 drivers/iommu/intel-iommu.c if (!dma_pte_present(pte)) { pte 956 drivers/iommu/intel-iommu.c if (dma_pte_superpage(pte)) { pte 958 drivers/iommu/intel-iommu.c return pte; pte 961 drivers/iommu/intel-iommu.c parent = phys_to_virt(dma_pte_addr(pte)); pte 973 drivers/iommu/intel-iommu.c struct dma_pte *first_pte, *pte; pte 982 drivers/iommu/intel-iommu.c first_pte = pte = dma_pfn_level_pte(domain, start_pfn, 1, &large_page); pte 983 drivers/iommu/intel-iommu.c if (!pte) { pte 988 drivers/iommu/intel-iommu.c dma_clear_pte(pte); pte 990 drivers/iommu/intel-iommu.c pte++; pte 991 drivers/iommu/intel-iommu.c } while (start_pfn <= last_pfn && !first_pte_in_page(pte)); pte 994 drivers/iommu/intel-iommu.c (void *)pte - (void *)first_pte); pte 1000 drivers/iommu/intel-iommu.c int retain_level, struct dma_pte *pte, pte 1005 drivers/iommu/intel-iommu.c pte = &pte[pfn_level_offset(pfn, level)]; pte 1011 drivers/iommu/intel-iommu.c if (!dma_pte_present(pte) || dma_pte_superpage(pte)) pte 1015 drivers/iommu/intel-iommu.c level_pte = phys_to_virt(dma_pte_addr(pte)); pte 1029 drivers/iommu/intel-iommu.c dma_clear_pte(pte); pte 1030 drivers/iommu/intel-iommu.c domain_flush_cache(domain, pte, sizeof(*pte)); pte 1035 drivers/iommu/intel-iommu.c } while (!first_pte_in_page(++pte) && pfn <= last_pfn); pte 1071 drivers/iommu/intel-iommu.c int level, struct dma_pte *pte, pte 1076 drivers/iommu/intel-iommu.c pg = pfn_to_page(dma_pte_addr(pte) >> PAGE_SHIFT); pte 1083 drivers/iommu/intel-iommu.c pte = page_address(pg); pte 1085 drivers/iommu/intel-iommu.c if (dma_pte_present(pte) && !dma_pte_superpage(pte)) pte 1087 drivers/iommu/intel-iommu.c pte, freelist); pte 1088 drivers/iommu/intel-iommu.c pte++; pte 1089 drivers/iommu/intel-iommu.c } while (!first_pte_in_page(pte)); pte 1095 drivers/iommu/intel-iommu.c struct dma_pte *pte, unsigned long pfn, pte 1103 drivers/iommu/intel-iommu.c pte = &pte[pfn_level_offset(pfn, level)]; pte 1108 drivers/iommu/intel-iommu.c if (!dma_pte_present(pte)) pte 1118 drivers/iommu/intel-iommu.c if (level > 1 && !dma_pte_superpage(pte)) pte 1119 drivers/iommu/intel-iommu.c freelist = dma_pte_list_pagetables(domain, level - 1, pte, freelist); pte 1121 drivers/iommu/intel-iommu.c dma_clear_pte(pte); pte 1123 drivers/iommu/intel-iommu.c first_pte = pte; pte 1124 drivers/iommu/intel-iommu.c last_pte = pte; pte 1128 drivers/iommu/intel-iommu.c phys_to_virt(dma_pte_addr(pte)), pte 1134 drivers/iommu/intel-iommu.c } while (!first_pte_in_page(++pte) && pfn <= last_pfn); pte 2227 drivers/iommu/intel-iommu.c struct dma_pte *first_pte = NULL, *pte = NULL; pte 2258 drivers/iommu/intel-iommu.c if (!pte) { pte 2261 drivers/iommu/intel-iommu.c first_pte = pte = pfn_to_dma_pte(domain, iov_pfn, &largepage_lvl); pte 2262 drivers/iommu/intel-iommu.c if (!pte) pte 2290 drivers/iommu/intel-iommu.c tmp = cmpxchg64_local(&pte->val, 0ULL, pteval); pte 2324 drivers/iommu/intel-iommu.c pte++; pte 2325 drivers/iommu/intel-iommu.c if (!nr_pages || first_pte_in_page(pte) || pte 2328 drivers/iommu/intel-iommu.c (void *)pte - (void *)first_pte); pte 2329 drivers/iommu/intel-iommu.c pte = NULL; pte 5368 drivers/iommu/intel-iommu.c struct dma_pte *pte; pte 5370 drivers/iommu/intel-iommu.c pte = dmar_domain->pgd; pte 5371 drivers/iommu/intel-iommu.c if (dma_pte_present(pte)) { pte 5373 drivers/iommu/intel-iommu.c phys_to_virt(dma_pte_addr(pte)); pte 5374 drivers/iommu/intel-iommu.c free_pgtable_page(pte); pte 5517 drivers/iommu/intel-iommu.c struct dma_pte *pte; pte 5521 drivers/iommu/intel-iommu.c pte = pfn_to_dma_pte(dmar_domain, iova >> VTD_PAGE_SHIFT, &level); pte 5522 drivers/iommu/intel-iommu.c if (pte && dma_pte_present(pte)) pte 5523 drivers/iommu/intel-iommu.c phys = dma_pte_addr(pte) + pte 447 drivers/iommu/intel-pasid.c struct pasid_entry *pte; pte 450 drivers/iommu/intel-pasid.c pte = intel_pasid_get_entry(dev, pasid); pte 451 drivers/iommu/intel-pasid.c if (WARN_ON(!pte)) pte 454 drivers/iommu/intel-pasid.c did = pasid_get_domain_id(pte); pte 458 drivers/iommu/intel-pasid.c clflush_cache_range(pte, sizeof(*pte)); pte 476 drivers/iommu/intel-pasid.c struct pasid_entry *pte; pte 484 drivers/iommu/intel-pasid.c pte = intel_pasid_get_entry(dev, pasid); pte 485 drivers/iommu/intel-pasid.c if (WARN_ON(!pte)) pte 488 drivers/iommu/intel-pasid.c pasid_clear_entry(pte); pte 491 drivers/iommu/intel-pasid.c pasid_set_flptr(pte, (u64)__pa(pgd)); pte 498 drivers/iommu/intel-pasid.c pasid_set_sre(pte); pte 505 drivers/iommu/intel-pasid.c pasid_set_flpm(pte, 1); pte 508 drivers/iommu/intel-pasid.c pasid_clear_entry(pte); pte 514 drivers/iommu/intel-pasid.c pasid_set_domain_id(pte, did); pte 515 drivers/iommu/intel-pasid.c pasid_set_address_width(pte, iommu->agaw); pte 516 drivers/iommu/intel-pasid.c pasid_set_page_snoop(pte, !!ecap_smpwc(iommu->ecap)); pte 519 drivers/iommu/intel-pasid.c pasid_set_translation_type(pte, 1); pte 520 drivers/iommu/intel-pasid.c pasid_set_present(pte); pte 523 drivers/iommu/intel-pasid.c clflush_cache_range(pte, sizeof(*pte)); pte 542 drivers/iommu/intel-pasid.c struct pasid_entry *pte; pte 574 drivers/iommu/intel-pasid.c pte = intel_pasid_get_entry(dev, pasid); pte 575 drivers/iommu/intel-pasid.c if (!pte) { pte 580 drivers/iommu/intel-pasid.c pasid_clear_entry(pte); pte 581 drivers/iommu/intel-pasid.c pasid_set_domain_id(pte, did); pte 582 drivers/iommu/intel-pasid.c pasid_set_slptr(pte, pgd_val); pte 583 drivers/iommu/intel-pasid.c pasid_set_address_width(pte, agaw); pte 584 drivers/iommu/intel-pasid.c pasid_set_translation_type(pte, 2); pte 585 drivers/iommu/intel-pasid.c pasid_set_fault_enable(pte); pte 586 drivers/iommu/intel-pasid.c pasid_set_page_snoop(pte, !!ecap_smpwc(iommu->ecap)); pte 592 drivers/iommu/intel-pasid.c pasid_set_sre(pte); pte 593 drivers/iommu/intel-pasid.c pasid_set_present(pte); pte 596 drivers/iommu/intel-pasid.c clflush_cache_range(pte, sizeof(*pte)); pte 616 drivers/iommu/intel-pasid.c struct pasid_entry *pte; pte 618 drivers/iommu/intel-pasid.c pte = intel_pasid_get_entry(dev, pasid); pte 619 drivers/iommu/intel-pasid.c if (!pte) { pte 624 drivers/iommu/intel-pasid.c pasid_clear_entry(pte); pte 625 drivers/iommu/intel-pasid.c pasid_set_domain_id(pte, did); pte 626 drivers/iommu/intel-pasid.c pasid_set_address_width(pte, iommu->agaw); pte 627 drivers/iommu/intel-pasid.c pasid_set_translation_type(pte, 4); pte 628 drivers/iommu/intel-pasid.c pasid_set_fault_enable(pte); pte 629 drivers/iommu/intel-pasid.c pasid_set_page_snoop(pte, !!ecap_smpwc(iommu->ecap)); pte 635 drivers/iommu/intel-pasid.c pasid_set_sre(pte); pte 636 drivers/iommu/intel-pasid.c pasid_set_present(pte); pte 639 drivers/iommu/intel-pasid.c clflush_cache_range(pte, sizeof(*pte)); pte 73 drivers/iommu/intel-pasid.h static inline bool pasid_pte_is_present(struct pasid_entry *pte) pte 75 drivers/iommu/intel-pasid.h return READ_ONCE(pte->val[0]) & PASID_PTE_PRESENT; pte 84 drivers/iommu/io-pgtable-arm-v7s.c #define ARM_V7S_PTE_IS_VALID(pte) (((pte) & 0x3) != 0) pte 85 drivers/iommu/io-pgtable-arm-v7s.c #define ARM_V7S_PTE_IS_TABLE(pte, lvl) \ pte 86 drivers/iommu/io-pgtable-arm-v7s.c ((lvl) == 1 && (((pte) & 0x3) == ARM_V7S_PTE_TYPE_TABLE)) pte 174 drivers/iommu/io-pgtable-arm-v7s.c static bool arm_v7s_pte_is_cont(arm_v7s_iopte pte, int lvl); pte 190 drivers/iommu/io-pgtable-arm-v7s.c arm_v7s_iopte pte = paddr & ARM_V7S_LVL_MASK(lvl); pte 193 drivers/iommu/io-pgtable-arm-v7s.c return pte; pte 196 drivers/iommu/io-pgtable-arm-v7s.c pte |= ARM_V7S_ATTR_MTK_PA_BIT32; pte 198 drivers/iommu/io-pgtable-arm-v7s.c pte |= ARM_V7S_ATTR_MTK_PA_BIT33; pte 199 drivers/iommu/io-pgtable-arm-v7s.c return pte; pte 202 drivers/iommu/io-pgtable-arm-v7s.c static phys_addr_t iopte_to_paddr(arm_v7s_iopte pte, int lvl, pte 208 drivers/iommu/io-pgtable-arm-v7s.c if (ARM_V7S_PTE_IS_TABLE(pte, lvl)) pte 210 drivers/iommu/io-pgtable-arm-v7s.c else if (arm_v7s_pte_is_cont(pte, lvl)) pte 215 drivers/iommu/io-pgtable-arm-v7s.c paddr = pte & mask; pte 219 drivers/iommu/io-pgtable-arm-v7s.c if (pte & ARM_V7S_ATTR_MTK_PA_BIT32) pte 221 drivers/iommu/io-pgtable-arm-v7s.c if (pte & ARM_V7S_ATTR_MTK_PA_BIT33) pte 226 drivers/iommu/io-pgtable-arm-v7s.c static arm_v7s_iopte *iopte_deref(arm_v7s_iopte pte, int lvl, pte 229 drivers/iommu/io-pgtable-arm-v7s.c return phys_to_virt(iopte_to_paddr(pte, lvl, &data->iop.cfg)); pte 305 drivers/iommu/io-pgtable-arm-v7s.c static void __arm_v7s_set_pte(arm_v7s_iopte *ptep, arm_v7s_iopte pte, pte 311 drivers/iommu/io-pgtable-arm-v7s.c ptep[i] = pte; pte 320 drivers/iommu/io-pgtable-arm-v7s.c arm_v7s_iopte pte = ARM_V7S_ATTR_NG | ARM_V7S_ATTR_S; pte 323 drivers/iommu/io-pgtable-arm-v7s.c pte |= ARM_V7S_ATTR_TEX(1); pte 325 drivers/iommu/io-pgtable-arm-v7s.c pte |= ARM_V7S_PTE_AF; pte 327 drivers/iommu/io-pgtable-arm-v7s.c pte |= ARM_V7S_PTE_AP_UNPRIV; pte 329 drivers/iommu/io-pgtable-arm-v7s.c pte |= ARM_V7S_PTE_AP_RDONLY; pte 331 drivers/iommu/io-pgtable-arm-v7s.c pte <<= ARM_V7S_ATTR_SHIFT(lvl); pte 334 drivers/iommu/io-pgtable-arm-v7s.c pte |= ARM_V7S_ATTR_XN(lvl); pte 336 drivers/iommu/io-pgtable-arm-v7s.c pte |= ARM_V7S_ATTR_B; pte 338 drivers/iommu/io-pgtable-arm-v7s.c pte |= ARM_V7S_ATTR_B | ARM_V7S_ATTR_C; pte 340 drivers/iommu/io-pgtable-arm-v7s.c pte |= ARM_V7S_PTE_TYPE_PAGE; pte 342 drivers/iommu/io-pgtable-arm-v7s.c pte |= ARM_V7S_ATTR_NS_SECTION; pte 344 drivers/iommu/io-pgtable-arm-v7s.c return pte; pte 347 drivers/iommu/io-pgtable-arm-v7s.c static int arm_v7s_pte_to_prot(arm_v7s_iopte pte, int lvl) pte 350 drivers/iommu/io-pgtable-arm-v7s.c arm_v7s_iopte attr = pte >> ARM_V7S_ATTR_SHIFT(lvl); pte 358 drivers/iommu/io-pgtable-arm-v7s.c else if (pte & ARM_V7S_ATTR_C) pte 360 drivers/iommu/io-pgtable-arm-v7s.c if (pte & ARM_V7S_ATTR_XN(lvl)) pte 366 drivers/iommu/io-pgtable-arm-v7s.c static arm_v7s_iopte arm_v7s_pte_to_cont(arm_v7s_iopte pte, int lvl) pte 369 drivers/iommu/io-pgtable-arm-v7s.c pte |= ARM_V7S_CONT_SECTION; pte 371 drivers/iommu/io-pgtable-arm-v7s.c arm_v7s_iopte xn = pte & ARM_V7S_ATTR_XN(lvl); pte 372 drivers/iommu/io-pgtable-arm-v7s.c arm_v7s_iopte tex = pte & ARM_V7S_CONT_PAGE_TEX_MASK; pte 374 drivers/iommu/io-pgtable-arm-v7s.c pte ^= xn | tex | ARM_V7S_PTE_TYPE_PAGE; pte 375 drivers/iommu/io-pgtable-arm-v7s.c pte |= (xn << ARM_V7S_CONT_PAGE_XN_SHIFT) | pte 379 drivers/iommu/io-pgtable-arm-v7s.c return pte; pte 382 drivers/iommu/io-pgtable-arm-v7s.c static arm_v7s_iopte arm_v7s_cont_to_pte(arm_v7s_iopte pte, int lvl) pte 385 drivers/iommu/io-pgtable-arm-v7s.c pte &= ~ARM_V7S_CONT_SECTION; pte 387 drivers/iommu/io-pgtable-arm-v7s.c arm_v7s_iopte xn = pte & BIT(ARM_V7S_CONT_PAGE_XN_SHIFT); pte 388 drivers/iommu/io-pgtable-arm-v7s.c arm_v7s_iopte tex = pte & (ARM_V7S_CONT_PAGE_TEX_MASK << pte 391 drivers/iommu/io-pgtable-arm-v7s.c pte ^= xn | tex | ARM_V7S_PTE_TYPE_CONT_PAGE; pte 392 drivers/iommu/io-pgtable-arm-v7s.c pte |= (xn >> ARM_V7S_CONT_PAGE_XN_SHIFT) | pte 396 drivers/iommu/io-pgtable-arm-v7s.c return pte; pte 399 drivers/iommu/io-pgtable-arm-v7s.c static bool arm_v7s_pte_is_cont(arm_v7s_iopte pte, int lvl) pte 401 drivers/iommu/io-pgtable-arm-v7s.c if (lvl == 1 && !ARM_V7S_PTE_IS_TABLE(pte, lvl)) pte 402 drivers/iommu/io-pgtable-arm-v7s.c return pte & ARM_V7S_CONT_SECTION; pte 404 drivers/iommu/io-pgtable-arm-v7s.c return !(pte & ARM_V7S_PTE_TYPE_PAGE); pte 417 drivers/iommu/io-pgtable-arm-v7s.c arm_v7s_iopte pte; pte 439 drivers/iommu/io-pgtable-arm-v7s.c pte = arm_v7s_prot_to_pte(prot, lvl, cfg); pte 441 drivers/iommu/io-pgtable-arm-v7s.c pte = arm_v7s_pte_to_cont(pte, lvl); pte 443 drivers/iommu/io-pgtable-arm-v7s.c pte |= paddr_to_iopte(paddr, lvl, cfg); pte 445 drivers/iommu/io-pgtable-arm-v7s.c __arm_v7s_set_pte(ptep, pte, num_entries, cfg); pte 478 drivers/iommu/io-pgtable-arm-v7s.c arm_v7s_iopte pte, *cptep; pte 494 drivers/iommu/io-pgtable-arm-v7s.c pte = READ_ONCE(*ptep); pte 495 drivers/iommu/io-pgtable-arm-v7s.c if (!pte) { pte 500 drivers/iommu/io-pgtable-arm-v7s.c pte = arm_v7s_install_table(cptep, ptep, 0, cfg); pte 501 drivers/iommu/io-pgtable-arm-v7s.c if (pte) pte 508 drivers/iommu/io-pgtable-arm-v7s.c if (ARM_V7S_PTE_IS_TABLE(pte, lvl)) { pte 509 drivers/iommu/io-pgtable-arm-v7s.c cptep = iopte_deref(pte, lvl, data); pte 510 drivers/iommu/io-pgtable-arm-v7s.c } else if (pte) { pte 556 drivers/iommu/io-pgtable-arm-v7s.c arm_v7s_iopte pte = data->pgd[i]; pte 558 drivers/iommu/io-pgtable-arm-v7s.c if (ARM_V7S_PTE_IS_TABLE(pte, 1)) pte 559 drivers/iommu/io-pgtable-arm-v7s.c __arm_v7s_free_table(iopte_deref(pte, 1, data), pte 572 drivers/iommu/io-pgtable-arm-v7s.c arm_v7s_iopte pte; pte 577 drivers/iommu/io-pgtable-arm-v7s.c pte = *ptep; pte 578 drivers/iommu/io-pgtable-arm-v7s.c if (!arm_v7s_pte_is_cont(pte, lvl)) pte 579 drivers/iommu/io-pgtable-arm-v7s.c return pte; pte 582 drivers/iommu/io-pgtable-arm-v7s.c pte = arm_v7s_cont_to_pte(pte, lvl); pte 584 drivers/iommu/io-pgtable-arm-v7s.c ptep[i] = pte + i * size; pte 590 drivers/iommu/io-pgtable-arm-v7s.c return pte; pte 600 drivers/iommu/io-pgtable-arm-v7s.c arm_v7s_iopte pte, *tablep; pte 611 drivers/iommu/io-pgtable-arm-v7s.c pte = arm_v7s_prot_to_pte(arm_v7s_pte_to_prot(blk_pte, 1), 2, cfg); pte 613 drivers/iommu/io-pgtable-arm-v7s.c pte = arm_v7s_pte_to_cont(pte, 2); pte 615 drivers/iommu/io-pgtable-arm-v7s.c for (i = 0; i < num_ptes; i += num_entries, pte += size) { pte 620 drivers/iommu/io-pgtable-arm-v7s.c __arm_v7s_set_pte(&tablep[i], pte, num_entries, cfg); pte 623 drivers/iommu/io-pgtable-arm-v7s.c pte = arm_v7s_install_table(tablep, ptep, blk_pte, cfg); pte 624 drivers/iommu/io-pgtable-arm-v7s.c if (pte != blk_pte) { pte 627 drivers/iommu/io-pgtable-arm-v7s.c if (!ARM_V7S_PTE_IS_TABLE(pte, 1)) pte 630 drivers/iommu/io-pgtable-arm-v7s.c tablep = iopte_deref(pte, 1, data); pte 643 drivers/iommu/io-pgtable-arm-v7s.c arm_v7s_iopte pte[ARM_V7S_CONT_PAGES]; pte 654 drivers/iommu/io-pgtable-arm-v7s.c pte[i] = READ_ONCE(ptep[i]); pte 655 drivers/iommu/io-pgtable-arm-v7s.c if (WARN_ON(!ARM_V7S_PTE_IS_VALID(pte[i]))) pte 669 drivers/iommu/io-pgtable-arm-v7s.c if (num_entries <= 1 && arm_v7s_pte_is_cont(pte[0], lvl)) { pte 673 drivers/iommu/io-pgtable-arm-v7s.c pte[0] = arm_v7s_split_cont(data, iova, idx, lvl, ptep); pte 684 drivers/iommu/io-pgtable-arm-v7s.c if (ARM_V7S_PTE_IS_TABLE(pte[i], lvl)) { pte 688 drivers/iommu/io-pgtable-arm-v7s.c ptep = iopte_deref(pte[i], lvl, data); pte 703 drivers/iommu/io-pgtable-arm-v7s.c } else if (lvl == 1 && !ARM_V7S_PTE_IS_TABLE(pte[0], lvl)) { pte 708 drivers/iommu/io-pgtable-arm-v7s.c return arm_v7s_split_blk_unmap(data, gather, iova, size, pte[0], pte 713 drivers/iommu/io-pgtable-arm-v7s.c ptep = iopte_deref(pte[0], lvl, data); pte 732 drivers/iommu/io-pgtable-arm-v7s.c arm_v7s_iopte *ptep = data->pgd, pte; pte 738 drivers/iommu/io-pgtable-arm-v7s.c pte = READ_ONCE(*ptep); pte 739 drivers/iommu/io-pgtable-arm-v7s.c ptep = iopte_deref(pte, lvl, data); pte 740 drivers/iommu/io-pgtable-arm-v7s.c } while (ARM_V7S_PTE_IS_TABLE(pte, lvl)); pte 742 drivers/iommu/io-pgtable-arm-v7s.c if (!ARM_V7S_PTE_IS_VALID(pte)) pte 746 drivers/iommu/io-pgtable-arm-v7s.c if (arm_v7s_pte_is_cont(pte, lvl)) pte 748 drivers/iommu/io-pgtable-arm-v7s.c return iopte_to_paddr(pte, lvl, &data->iop.cfg) | (iova & ~mask); pte 173 drivers/iommu/io-pgtable-arm.c #define iopte_deref(pte,d) __va(iopte_to_paddr(pte, d)) pte 175 drivers/iommu/io-pgtable-arm.c #define iopte_type(pte,l) \ pte 176 drivers/iommu/io-pgtable-arm.c (((pte) >> ARM_LPAE_PTE_TYPE_SHIFT) & ARM_LPAE_PTE_TYPE_MASK) pte 178 drivers/iommu/io-pgtable-arm.c #define iopte_prot(pte) ((pte) & ARM_LPAE_PTE_ATTR_MASK) pte 193 drivers/iommu/io-pgtable-arm.c static inline bool iopte_leaf(arm_lpae_iopte pte, int lvl, pte 197 drivers/iommu/io-pgtable-arm.c return iopte_type(pte, lvl) == ARM_LPAE_PTE_TYPE_PAGE; pte 199 drivers/iommu/io-pgtable-arm.c return iopte_type(pte, lvl) == ARM_LPAE_PTE_TYPE_BLOCK; pte 205 drivers/iommu/io-pgtable-arm.c arm_lpae_iopte pte = paddr; pte 208 drivers/iommu/io-pgtable-arm.c return (pte | (pte >> (48 - 12))) & ARM_LPAE_PTE_ADDR_MASK; pte 211 drivers/iommu/io-pgtable-arm.c static phys_addr_t iopte_to_paddr(arm_lpae_iopte pte, pte 214 drivers/iommu/io-pgtable-arm.c u64 paddr = pte & ARM_LPAE_PTE_ADDR_MASK; pte 285 drivers/iommu/io-pgtable-arm.c static void __arm_lpae_set_pte(arm_lpae_iopte *ptep, arm_lpae_iopte pte, pte 288 drivers/iommu/io-pgtable-arm.c *ptep = pte; pte 303 drivers/iommu/io-pgtable-arm.c arm_lpae_iopte pte = prot; pte 306 drivers/iommu/io-pgtable-arm.c pte |= ARM_LPAE_PTE_NS; pte 309 drivers/iommu/io-pgtable-arm.c pte |= ARM_LPAE_PTE_TYPE_PAGE; pte 311 drivers/iommu/io-pgtable-arm.c pte |= ARM_LPAE_PTE_TYPE_BLOCK; pte 314 drivers/iommu/io-pgtable-arm.c pte |= ARM_LPAE_PTE_AF; pte 315 drivers/iommu/io-pgtable-arm.c pte |= ARM_LPAE_PTE_SH_IS; pte 316 drivers/iommu/io-pgtable-arm.c pte |= paddr_to_iopte(paddr, data); pte 318 drivers/iommu/io-pgtable-arm.c __arm_lpae_set_pte(ptep, pte, &data->iop.cfg); pte 326 drivers/iommu/io-pgtable-arm.c arm_lpae_iopte pte = *ptep; pte 328 drivers/iommu/io-pgtable-arm.c if (iopte_leaf(pte, lvl, data->iop.fmt)) { pte 332 drivers/iommu/io-pgtable-arm.c } else if (iopte_type(pte, lvl) == ARM_LPAE_PTE_TYPE_TABLE) { pte 386 drivers/iommu/io-pgtable-arm.c arm_lpae_iopte *cptep, pte; pte 403 drivers/iommu/io-pgtable-arm.c pte = READ_ONCE(*ptep); pte 404 drivers/iommu/io-pgtable-arm.c if (!pte) { pte 409 drivers/iommu/io-pgtable-arm.c pte = arm_lpae_install_table(cptep, ptep, 0, cfg); pte 410 drivers/iommu/io-pgtable-arm.c if (pte) pte 412 drivers/iommu/io-pgtable-arm.c } else if (!cfg->coherent_walk && !(pte & ARM_LPAE_PTE_SW_SYNC)) { pte 416 drivers/iommu/io-pgtable-arm.c if (pte && !iopte_leaf(pte, lvl, data->iop.fmt)) { pte 417 drivers/iommu/io-pgtable-arm.c cptep = iopte_deref(pte, data); pte 418 drivers/iommu/io-pgtable-arm.c } else if (pte) { pte 431 drivers/iommu/io-pgtable-arm.c arm_lpae_iopte pte; pte 435 drivers/iommu/io-pgtable-arm.c pte = ARM_LPAE_PTE_nG; pte 437 drivers/iommu/io-pgtable-arm.c pte |= ARM_LPAE_PTE_AP_RDONLY; pte 439 drivers/iommu/io-pgtable-arm.c pte |= ARM_LPAE_PTE_AP_UNPRIV; pte 441 drivers/iommu/io-pgtable-arm.c pte = ARM_LPAE_PTE_HAP_FAULT; pte 443 drivers/iommu/io-pgtable-arm.c pte |= ARM_LPAE_PTE_HAP_READ; pte 445 drivers/iommu/io-pgtable-arm.c pte |= ARM_LPAE_PTE_HAP_WRITE; pte 455 drivers/iommu/io-pgtable-arm.c pte |= ARM_LPAE_PTE_MEMATTR_DEV; pte 457 drivers/iommu/io-pgtable-arm.c pte |= ARM_LPAE_PTE_MEMATTR_OIWB; pte 459 drivers/iommu/io-pgtable-arm.c pte |= ARM_LPAE_PTE_MEMATTR_NC; pte 462 drivers/iommu/io-pgtable-arm.c pte |= (ARM_LPAE_MAIR_ATTR_IDX_DEV pte 465 drivers/iommu/io-pgtable-arm.c pte |= (ARM_LPAE_MAIR_ATTR_IDX_CACHE pte 468 drivers/iommu/io-pgtable-arm.c pte |= (ARM_LPAE_MAIR_ATTR_IDX_INC_OCACHE pte 473 drivers/iommu/io-pgtable-arm.c pte |= ARM_LPAE_PTE_XN; pte 475 drivers/iommu/io-pgtable-arm.c return pte; pte 525 drivers/iommu/io-pgtable-arm.c arm_lpae_iopte pte = *ptep++; pte 527 drivers/iommu/io-pgtable-arm.c if (!pte || iopte_leaf(pte, lvl, data->iop.fmt)) pte 530 drivers/iommu/io-pgtable-arm.c __arm_lpae_free_pgtable(data, lvl + 1, iopte_deref(pte, data)); pte 551 drivers/iommu/io-pgtable-arm.c arm_lpae_iopte pte, *tablep; pte 568 drivers/iommu/io-pgtable-arm.c pte = iopte_prot(blk_pte); pte 570 drivers/iommu/io-pgtable-arm.c for (i = 0; i < tablesz / sizeof(pte); i++, blk_paddr += split_sz) { pte 575 drivers/iommu/io-pgtable-arm.c __arm_lpae_init_pte(data, blk_paddr, pte, lvl, &tablep[i]); pte 578 drivers/iommu/io-pgtable-arm.c pte = arm_lpae_install_table(tablep, ptep, blk_pte, cfg); pte 579 drivers/iommu/io-pgtable-arm.c if (pte != blk_pte) { pte 586 drivers/iommu/io-pgtable-arm.c if (iopte_type(pte, lvl - 1) != ARM_LPAE_PTE_TYPE_TABLE) pte 589 drivers/iommu/io-pgtable-arm.c tablep = iopte_deref(pte, data); pte 603 drivers/iommu/io-pgtable-arm.c arm_lpae_iopte pte; pte 611 drivers/iommu/io-pgtable-arm.c pte = READ_ONCE(*ptep); pte 612 drivers/iommu/io-pgtable-arm.c if (WARN_ON(!pte)) pte 619 drivers/iommu/io-pgtable-arm.c if (!iopte_leaf(pte, lvl, iop->fmt)) { pte 623 drivers/iommu/io-pgtable-arm.c ptep = iopte_deref(pte, data); pte 637 drivers/iommu/io-pgtable-arm.c } else if (iopte_leaf(pte, lvl, iop->fmt)) { pte 642 drivers/iommu/io-pgtable-arm.c return arm_lpae_split_blk_unmap(data, gather, iova, size, pte, pte 647 drivers/iommu/io-pgtable-arm.c ptep = iopte_deref(pte, data); pte 668 drivers/iommu/io-pgtable-arm.c arm_lpae_iopte pte, *ptep = data->pgd; pte 678 drivers/iommu/io-pgtable-arm.c pte = READ_ONCE(*ptep); pte 681 drivers/iommu/io-pgtable-arm.c if (!pte) pte 685 drivers/iommu/io-pgtable-arm.c if (iopte_leaf(pte, lvl, data->iop.fmt)) pte 689 drivers/iommu/io-pgtable-arm.c ptep = iopte_deref(pte, data); pte 697 drivers/iommu/io-pgtable-arm.c return iopte_to_paddr(pte, data) | iova; pte 1630 drivers/iommu/omap-iommu.c u32 *pgd, *pte; pte 1637 drivers/iommu/omap-iommu.c iopgtable_lookup_entry(oiommu, da, &pgd, &pte); pte 1639 drivers/iommu/omap-iommu.c if (pte) { pte 1640 drivers/iommu/omap-iommu.c if (iopte_is_small(*pte)) pte 1641 drivers/iommu/omap-iommu.c ret = omap_iommu_translate(*pte, da, IOPTE_MASK); pte 1642 drivers/iommu/omap-iommu.c else if (iopte_is_large(*pte)) pte 1643 drivers/iommu/omap-iommu.c ret = omap_iommu_translate(*pte, da, IOLARGE_MASK); pte 1645 drivers/iommu/omap-iommu.c dev_err(dev, "bogus pte 0x%x, da 0x%llx", *pte, pte 218 drivers/iommu/rockchip-iommu.c static inline phys_addr_t rk_pte_page_address(u32 pte) pte 220 drivers/iommu/rockchip-iommu.c return (phys_addr_t)pte & RK_PTE_PAGE_ADDRESS_MASK; pte 223 drivers/iommu/rockchip-iommu.c static inline bool rk_pte_is_page_valid(u32 pte) pte 225 drivers/iommu/rockchip-iommu.c return pte & RK_PTE_PAGE_VALID; pte 238 drivers/iommu/rockchip-iommu.c static u32 rk_mk_pte_invalid(u32 pte) pte 240 drivers/iommu/rockchip-iommu.c return pte & ~RK_PTE_PAGE_VALID; pte 483 drivers/iommu/rockchip-iommu.c u32 pte = 0; pte 503 drivers/iommu/rockchip-iommu.c pte = *pte_addr; pte 505 drivers/iommu/rockchip-iommu.c if (!rk_pte_is_page_valid(pte)) pte 508 drivers/iommu/rockchip-iommu.c page_addr_phys = rk_pte_page_address(pte) + page_offset; pte 509 drivers/iommu/rockchip-iommu.c page_flags = pte & RK_PTE_PAGE_FLAGS_MASK; pte 516 drivers/iommu/rockchip-iommu.c rk_dte_is_pt_valid(dte), &pte_addr_phys, pte, pte 517 drivers/iommu/rockchip-iommu.c rk_pte_is_page_valid(pte), &page_addr_phys, page_flags); pte 595 drivers/iommu/rockchip-iommu.c u32 dte, pte; pte 606 drivers/iommu/rockchip-iommu.c pte = page_table[rk_iova_pte_index(iova)]; pte 607 drivers/iommu/rockchip-iommu.c if (!rk_pte_is_page_valid(pte)) pte 610 drivers/iommu/rockchip-iommu.c phys = rk_pte_page_address(pte) + rk_iova_page_offset(iova); pte 703 drivers/iommu/rockchip-iommu.c u32 pte = pte_addr[pte_count]; pte 704 drivers/iommu/rockchip-iommu.c if (!rk_pte_is_page_valid(pte)) pte 707 drivers/iommu/rockchip-iommu.c pte_addr[pte_count] = rk_mk_pte_invalid(pte); pte 726 drivers/iommu/rockchip-iommu.c u32 pte = pte_addr[pte_count]; pte 728 drivers/iommu/rockchip-iommu.c if (rk_pte_is_page_valid(pte)) pte 66 drivers/iommu/tegra-gart.c unsigned long iova, unsigned long pte) pte 69 drivers/iommu/tegra-gart.c writel_relaxed(pte, gart->regs + GART_ENTRY_DATA); pte 75 drivers/iommu/tegra-gart.c unsigned long pte; pte 78 drivers/iommu/tegra-gart.c pte = readl_relaxed(gart->regs + GART_ENTRY_DATA); pte 80 drivers/iommu/tegra-gart.c return pte; pte 229 drivers/iommu/tegra-gart.c unsigned long pte; pte 235 drivers/iommu/tegra-gart.c pte = gart_read_pte(gart, iova); pte 238 drivers/iommu/tegra-gart.c return pte & GART_PAGE_MASK; pte 639 drivers/iommu/tegra-smmu.c u32 *pte, dma_addr_t pte_dma, u32 val) pte 642 drivers/iommu/tegra-smmu.c unsigned long offset = offset_in_page(pte); pte 644 drivers/iommu/tegra-smmu.c *pte = val; pte 659 drivers/iommu/tegra-smmu.c u32 *pte; pte 661 drivers/iommu/tegra-smmu.c pte = as_get_pte(as, iova, &pte_dma); pte 662 drivers/iommu/tegra-smmu.c if (!pte) pte 666 drivers/iommu/tegra-smmu.c if (*pte == 0) pte 677 drivers/iommu/tegra-smmu.c tegra_smmu_set_pte(as, iova, pte, pte_dma, pte 688 drivers/iommu/tegra-smmu.c u32 *pte; pte 690 drivers/iommu/tegra-smmu.c pte = tegra_smmu_pte_lookup(as, iova, &pte_dma); pte 691 drivers/iommu/tegra-smmu.c if (!pte || !*pte) pte 694 drivers/iommu/tegra-smmu.c tegra_smmu_set_pte(as, iova, pte, pte_dma, 0); pte 706 drivers/iommu/tegra-smmu.c u32 *pte; pte 708 drivers/iommu/tegra-smmu.c pte = tegra_smmu_pte_lookup(as, iova, &pte_dma); pte 709 drivers/iommu/tegra-smmu.c if (!pte || !*pte) pte 712 drivers/iommu/tegra-smmu.c pfn = *pte & as->smmu->pfn_mask; pte 179 drivers/md/dm-switch.c region_table_slot_t pte; pte 183 drivers/md/dm-switch.c pte = sctx->region_table[region_index]; pte 184 drivers/md/dm-switch.c pte &= ~((((region_table_slot_t)1 << sctx->region_table_entry_bits) - 1) << bit); pte 185 drivers/md/dm-switch.c pte |= (region_table_slot_t)value << bit; pte 186 drivers/md/dm-switch.c sctx->region_table[region_index] = pte; pte 212 drivers/misc/sgi-gru/grufault.c pte_t pte; pte 231 drivers/misc/sgi-gru/grufault.c pte = *(pte_t *) pmdp; pte 234 drivers/misc/sgi-gru/grufault.c pte = *pte_offset_kernel(pmdp, vaddr); pte 236 drivers/misc/sgi-gru/grufault.c if (unlikely(!pte_present(pte) || pte 237 drivers/misc/sgi-gru/grufault.c (write && (!pte_write(pte) || !pte_dirty(pte))))) pte 240 drivers/misc/sgi-gru/grufault.c *paddr = pte_pfn(pte) << PAGE_SHIFT; pte 878 drivers/net/ethernet/mellanox/mlxsw/reg.h MLXSW_ITEM32(reg, spvm, pte, 0x00, 30, 1); pte 320 drivers/staging/gasket/gasket_page_table.c struct gasket_page_table_entry *pte, pte 324 drivers/staging/gasket/gasket_page_table.c pte->status = PTE_FREE; pte 329 drivers/staging/gasket/gasket_page_table.c if (pte->dma_addr) pte 330 drivers/staging/gasket/gasket_page_table.c dma_unmap_page(pg_tbl->device, pte->dma_addr, PAGE_SIZE, pte 333 drivers/staging/gasket/gasket_page_table.c vfree(pte->sublevel); pte 335 drivers/staging/gasket/gasket_page_table.c if (pte->page) pte 336 drivers/staging/gasket/gasket_page_table.c free_page((ulong)page_address(pte->page)); pte 338 drivers/staging/gasket/gasket_page_table.c memset(pte, 0, sizeof(struct gasket_page_table_entry)); pte 348 drivers/staging/gasket/gasket_page_table.c struct gasket_page_table_entry *pte; pte 354 drivers/staging/gasket/gasket_page_table.c for (pte = pg_tbl->entries + pg_tbl->num_simple_entries, pte 356 drivers/staging/gasket/gasket_page_table.c pte < pg_tbl->entries + pg_tbl->config.total_entries; pte 357 drivers/staging/gasket/gasket_page_table.c pte++, slot++) { pte 358 drivers/staging/gasket/gasket_page_table.c if (pte->status == PTE_INUSE) { pte 359 drivers/staging/gasket/gasket_page_table.c if (gasket_is_pte_range_free(pte->sublevel, pte 361 drivers/staging/gasket/gasket_page_table.c gasket_free_extended_subtable(pg_tbl, pte, pte 646 drivers/staging/gasket/gasket_page_table.c struct gasket_page_table_entry *pte; pte 651 drivers/staging/gasket/gasket_page_table.c pte = pg_tbl->entries + pg_tbl->num_simple_entries + pte 658 drivers/staging/gasket/gasket_page_table.c if (pte->status == PTE_INUSE) { pte 659 drivers/staging/gasket/gasket_page_table.c slot_base = (u64 __iomem *)(page_address(pte->page) + pte 660 drivers/staging/gasket/gasket_page_table.c pte->offset); pte 662 drivers/staging/gasket/gasket_page_table.c pte->sublevel + slot_idx, pte 668 drivers/staging/gasket/gasket_page_table.c pte++; pte 848 drivers/staging/gasket/gasket_page_table.c struct gasket_page_table_entry *pte, pte 863 drivers/staging/gasket/gasket_page_table.c pte->page = virt_to_page((void *)page_addr); pte 864 drivers/staging/gasket/gasket_page_table.c pte->offset = 0; pte 868 drivers/staging/gasket/gasket_page_table.c pte->sublevel = vzalloc(subtable_bytes); pte 869 drivers/staging/gasket/gasket_page_table.c if (!pte->sublevel) { pte 871 drivers/staging/gasket/gasket_page_table.c memset(pte, 0, sizeof(struct gasket_page_table_entry)); pte 876 drivers/staging/gasket/gasket_page_table.c pte->dma_addr = dma_map_page(pg_tbl->device, pte->page, 0, PAGE_SIZE, pte 878 drivers/staging/gasket/gasket_page_table.c if (dma_mapping_error(pg_tbl->device, pte->dma_addr)) { pte 880 drivers/staging/gasket/gasket_page_table.c vfree(pte->sublevel); pte 881 drivers/staging/gasket/gasket_page_table.c memset(pte, 0, sizeof(struct gasket_page_table_entry)); pte 886 drivers/staging/gasket/gasket_page_table.c dma_addr = (pte->dma_addr + pte->offset) | GASKET_VALID_SLOT_FLAG; pte 889 drivers/staging/gasket/gasket_page_table.c pte->status = PTE_INUSE; pte 913 drivers/staging/gasket/gasket_page_table.c struct gasket_page_table_entry *pte; pte 918 drivers/staging/gasket/gasket_page_table.c pte = pg_tbl->entries + pg_tbl->num_simple_entries + pte 927 drivers/staging/gasket/gasket_page_table.c if (pte->status == PTE_FREE) { pte 928 drivers/staging/gasket/gasket_page_table.c ret = gasket_alloc_extended_subtable(pg_tbl, pte, slot); pte 935 drivers/staging/gasket/gasket_page_table.c if (!gasket_is_pte_range_free(pte->sublevel + pte 942 drivers/staging/gasket/gasket_page_table.c pte++; pte 960 drivers/staging/gasket/gasket_page_table.c struct gasket_page_table_entry *pte; pte 978 drivers/staging/gasket/gasket_page_table.c pte = pg_tbl->entries + pg_tbl->num_simple_entries + pte 985 drivers/staging/gasket/gasket_page_table.c (u64 __iomem *)(page_address(pte->page) + pte->offset); pte 986 drivers/staging/gasket/gasket_page_table.c ret = gasket_perform_mapping(pg_tbl, pte->sublevel + slot_idx, pte 997 drivers/staging/gasket/gasket_page_table.c pte++; pte 1093 drivers/staging/gasket/gasket_page_table.c struct gasket_page_table_entry *pte; pte 1101 drivers/staging/gasket/gasket_page_table.c pte = pg_tbl->entries + page_num; pte 1102 drivers/staging/gasket/gasket_page_table.c if (pte->status != PTE_INUSE) pte 1110 drivers/staging/gasket/gasket_page_table.c pte = pg_tbl->entries + pg_tbl->num_simple_entries + page_num; pte 1111 drivers/staging/gasket/gasket_page_table.c if (pte->status != PTE_INUSE) pte 1116 drivers/staging/gasket/gasket_page_table.c pte = pte->sublevel + page_num; pte 1117 drivers/staging/gasket/gasket_page_table.c if (pte->status != PTE_INUSE) pte 1121 drivers/staging/gasket/gasket_page_table.c *ppage = pte->page; pte 1122 drivers/staging/gasket/gasket_page_table.c *poffset = pte->offset; pte 29 drivers/staging/media/ipu3/ipu3-mmu.c #define IPU3_PTE2ADDR(pte) ((phys_addr_t)(pte) << IPU3_PAGE_SHIFT) pte 124 drivers/staging/media/ipu3/ipu3-mmu.c int pte; pte 130 drivers/staging/media/ipu3/ipu3-mmu.c for (pte = 0; pte < IPU3_PT_PTES; pte++) pte 131 drivers/staging/media/ipu3/ipu3-mmu.c pt[pte] = pteval; pte 1247 drivers/target/sbp/sbp_target.c struct sbp_page_table_entry *pte; pte 1276 drivers/target/sbp/sbp_target.c pte = req->pg_tbl; pte 1282 drivers/target/sbp/sbp_target.c pte = NULL; pte 1294 drivers/target/sbp/sbp_target.c offset = (u64)be16_to_cpu(pte->segment_base_hi) << 32 | pte 1295 drivers/target/sbp/sbp_target.c be32_to_cpu(pte->segment_base_lo); pte 1296 drivers/target/sbp/sbp_target.c length = be16_to_cpu(pte->segment_length); pte 1298 drivers/target/sbp/sbp_target.c pte++; pte 265 drivers/xen/gntdev.c static int find_grant_ptes(pte_t *pte, unsigned long addr, void *data) pte 273 drivers/xen/gntdev.c pte_maddr = arbitrary_virt_to_machine(pte).maddr; pte 292 drivers/xen/gntdev.c static int set_grant_ptes_as_special(pte_t *pte, unsigned long addr, void *data) pte 294 drivers/xen/gntdev.c set_pte_at(current->mm, addr, pte, pte_mkspecial(*pte)); pte 943 drivers/xen/privcmd.c static int is_mapped_fn(pte_t *pte, unsigned long addr, void *data) pte 945 drivers/xen/privcmd.c return pte_none(*pte) ? 0 : -EBUSY; pte 100 drivers/xen/xlate_mmu.c pte_t pte = pte_mkspecial(pfn_pte(page_to_pfn(page), info->prot)); pte 138 drivers/xen/xlate_mmu.c set_pte_at(info->vma->vm_mm, addr, ptep, pte); pte 277 drivers/xen/xlate_mmu.c pte_t pte = pte_mkspecial(pfn_pte(page_to_pfn(page), r->prot)); pte 279 drivers/xen/xlate_mmu.c set_pte_at(r->mm, addr, ptep, pte); pte 780 fs/dax.c pte_t pte, *ptep = NULL; pte 836 fs/dax.c pte = ptep_clear_flush(vma, address, ptep); pte 837 fs/dax.c pte = pte_wrprotect(pte); pte 838 fs/dax.c pte = pte_mkclean(pte); pte 839 fs/dax.c set_pte_at(vma->vm_mm, address, ptep, pte); pte 521 fs/proc/task_mmu.c static void smaps_pte_entry(pte_t *pte, unsigned long addr, pte 529 fs/proc/task_mmu.c if (pte_present(*pte)) { pte 530 fs/proc/task_mmu.c page = vm_normal_page(vma, addr, *pte); pte 531 fs/proc/task_mmu.c } else if (is_swap_pte(*pte)) { pte 532 fs/proc/task_mmu.c swp_entry_t swpent = pte_to_swp_entry(*pte); pte 552 fs/proc/task_mmu.c && pte_none(*pte))) { pte 569 fs/proc/task_mmu.c smaps_account(mss, page, false, pte_young(*pte), pte_dirty(*pte), locked); pte 606 fs/proc/task_mmu.c pte_t *pte; pte 624 fs/proc/task_mmu.c pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl); pte 625 fs/proc/task_mmu.c for (; addr != end; pte++, addr += PAGE_SIZE) pte 626 fs/proc/task_mmu.c smaps_pte_entry(pte, addr, walk); pte 627 fs/proc/task_mmu.c pte_unmap_unlock(pte - 1, ptl); pte 707 fs/proc/task_mmu.c static int smaps_hugetlb_range(pte_t *pte, unsigned long hmask, pte 715 fs/proc/task_mmu.c if (pte_present(*pte)) { pte 716 fs/proc/task_mmu.c page = vm_normal_page(vma, addr, *pte); pte 717 fs/proc/task_mmu.c } else if (is_swap_pte(*pte)) { pte 718 fs/proc/task_mmu.c swp_entry_t swpent = pte_to_swp_entry(*pte); pte 989 fs/proc/task_mmu.c unsigned long addr, pte_t *pte) pte 997 fs/proc/task_mmu.c pte_t ptent = *pte; pte 1002 fs/proc/task_mmu.c old_pte = ptep_modify_prot_start(vma, addr, pte); pte 1005 fs/proc/task_mmu.c ptep_modify_prot_commit(vma, addr, pte, old_pte, ptent); pte 1008 fs/proc/task_mmu.c set_pte_at(vma->vm_mm, addr, pte, ptent); pte 1013 fs/proc/task_mmu.c unsigned long addr, pte_t *pte) pte 1053 fs/proc/task_mmu.c pte_t *pte, ptent; pte 1081 fs/proc/task_mmu.c pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl); pte 1082 fs/proc/task_mmu.c for (; addr != end; pte++, addr += PAGE_SIZE) { pte 1083 fs/proc/task_mmu.c ptent = *pte; pte 1086 fs/proc/task_mmu.c clear_soft_dirty(vma, addr, pte); pte 1098 fs/proc/task_mmu.c ptep_test_and_clear_young(vma, addr, pte); pte 1102 fs/proc/task_mmu.c pte_unmap_unlock(pte - 1, ptl); pte 1325 fs/proc/task_mmu.c struct vm_area_struct *vma, unsigned long addr, pte_t pte) pte 1330 fs/proc/task_mmu.c if (pte_present(pte)) { pte 1332 fs/proc/task_mmu.c frame = pte_pfn(pte); pte 1334 fs/proc/task_mmu.c page = vm_normal_page(vma, addr, pte); pte 1335 fs/proc/task_mmu.c if (pte_soft_dirty(pte)) pte 1337 fs/proc/task_mmu.c } else if (is_swap_pte(pte)) { pte 1339 fs/proc/task_mmu.c if (pte_swp_soft_dirty(pte)) pte 1341 fs/proc/task_mmu.c entry = pte_to_swp_entry(pte); pte 1369 fs/proc/task_mmu.c pte_t *pte, *orig_pte; pte 1439 fs/proc/task_mmu.c orig_pte = pte = pte_offset_map_lock(walk->mm, pmdp, addr, &ptl); pte 1440 fs/proc/task_mmu.c for (; addr < end; pte++, addr += PAGE_SIZE) { pte 1443 fs/proc/task_mmu.c pme = pte_to_pagemap_entry(pm, vma, addr, *pte); pte 1465 fs/proc/task_mmu.c pte_t pte; pte 1470 fs/proc/task_mmu.c pte = huge_ptep_get(ptep); pte 1471 fs/proc/task_mmu.c if (pte_present(pte)) { pte 1472 fs/proc/task_mmu.c struct page *page = pte_page(pte); pte 1482 fs/proc/task_mmu.c frame = pte_pfn(pte) + pte 1694 fs/proc/task_mmu.c static struct page *can_gather_numa_stats(pte_t pte, struct vm_area_struct *vma, pte 1700 fs/proc/task_mmu.c if (!pte_present(pte)) pte 1703 fs/proc/task_mmu.c page = vm_normal_page(vma, addr, pte); pte 1750 fs/proc/task_mmu.c pte_t *pte; pte 1768 fs/proc/task_mmu.c orig_pte = pte = pte_offset_map_lock(walk->mm, pmd, addr, &ptl); pte 1770 fs/proc/task_mmu.c struct page *page = can_gather_numa_stats(*pte, vma, addr); pte 1773 fs/proc/task_mmu.c gather_stats(page, md, pte_dirty(*pte), 1); pte 1775 fs/proc/task_mmu.c } while (pte++, addr += PAGE_SIZE, addr != end); pte 1781 fs/proc/task_mmu.c static int gather_hugetlb_stats(pte_t *pte, unsigned long hmask, pte 1784 fs/proc/task_mmu.c pte_t huge_pte = huge_ptep_get(pte); pte 1801 fs/proc/task_mmu.c static int gather_hugetlb_stats(pte_t *pte, unsigned long hmask, pte 234 fs/userfaultfd.c pte_t *ptep, pte; pte 245 fs/userfaultfd.c pte = huge_ptep_get(ptep); pte 251 fs/userfaultfd.c if (huge_pte_none(pte)) pte 253 fs/userfaultfd.c if (!huge_pte_write(pte) && (reason & VM_UFFD_WP)) pte 286 fs/userfaultfd.c pte_t *pte; pte 324 fs/userfaultfd.c pte = pte_offset_map(pmd, address); pte 329 fs/userfaultfd.c if (pte_none(*pte)) pte 331 fs/userfaultfd.c pte_unmap(pte); pte 10 include/asm-generic/hugetlb.h static inline unsigned long huge_pte_write(pte_t pte) pte 12 include/asm-generic/hugetlb.h return pte_write(pte); pte 15 include/asm-generic/hugetlb.h static inline unsigned long huge_pte_dirty(pte_t pte) pte 17 include/asm-generic/hugetlb.h return pte_dirty(pte); pte 20 include/asm-generic/hugetlb.h static inline pte_t huge_pte_mkwrite(pte_t pte) pte 22 include/asm-generic/hugetlb.h return pte_mkwrite(pte); pte 25 include/asm-generic/hugetlb.h static inline pte_t huge_pte_mkdirty(pte_t pte) pte 27 include/asm-generic/hugetlb.h return pte_mkdirty(pte); pte 30 include/asm-generic/hugetlb.h static inline pte_t huge_pte_modify(pte_t pte, pgprot_t newprot) pte 32 include/asm-generic/hugetlb.h return pte_modify(pte, newprot); pte 54 include/asm-generic/hugetlb.h pte_t *ptep, pte_t pte) pte 56 include/asm-generic/hugetlb.h set_pte_at(mm, addr, ptep, pte); pte 77 include/asm-generic/hugetlb.h static inline int huge_pte_none(pte_t pte) pte 79 include/asm-generic/hugetlb.h return pte_none(pte); pte 84 include/asm-generic/hugetlb.h static inline pte_t huge_pte_wrprotect(pte_t pte) pte 86 include/asm-generic/hugetlb.h return pte_wrprotect(pte); pte 116 include/asm-generic/hugetlb.h pte_t pte, int dirty) pte 118 include/asm-generic/hugetlb.h return ptep_set_access_flags(vma, addr, ptep, pte, dirty); pte 38 include/asm-generic/page.h unsigned long pte; pte 51 include/asm-generic/page.h #define pte_val(x) ((x).pte) pte 42 include/asm-generic/pgalloc.h static inline void pte_free_kernel(struct mm_struct *mm, pte_t *pte) pte 44 include/asm-generic/pgalloc.h free_page((unsigned long)pte); pte 61 include/asm-generic/pgalloc.h struct page *pte; pte 63 include/asm-generic/pgalloc.h pte = alloc_page(gfp); pte 64 include/asm-generic/pgalloc.h if (!pte) pte 66 include/asm-generic/pgalloc.h if (!pgtable_pte_page_ctor(pte)) { pte 67 include/asm-generic/pgalloc.h __free_page(pte); pte 71 include/asm-generic/pgalloc.h return pte; pte 36 include/asm-generic/pgtable-nopmd.h #define pud_populate(mm, pmd, pte) do { } while (0) pte 66 include/asm-generic/pgtable.h pte_t pte = *ptep; pte 68 include/asm-generic/pgtable.h if (!pte_young(pte)) pte 71 include/asm-generic/pgtable.h set_pte_at(vma->vm_mm, address, ptep, pte_mkold(pte)); pte 129 include/asm-generic/pgtable.h pte_t pte = *ptep; pte 131 include/asm-generic/pgtable.h return pte; pte 184 include/asm-generic/pgtable.h pte_t pte; pte 185 include/asm-generic/pgtable.h pte = ptep_get_and_clear(mm, address, ptep); pte 186 include/asm-generic/pgtable.h return pte; pte 346 include/asm-generic/pgtable.h static inline int pte_unused(pte_t pte) pte 353 include/asm-generic/pgtable.h #define pte_access_permitted(pte, write) \ pte 354 include/asm-generic/pgtable.h (pte_present(pte) && (!(write) || pte_write(pte))) pte 411 include/asm-generic/pgtable.h #define set_pte_safe(ptep, pte) \ pte 413 include/asm-generic/pgtable.h WARN_ON_ONCE(pte_present(*ptep) && !pte_same(*ptep, pte)); \ pte 414 include/asm-generic/pgtable.h set_pte(ptep, pte); \ pte 453 include/asm-generic/pgtable.h pte_t pte, pte_t oldpte) pte 482 include/asm-generic/pgtable.h #define move_pte(pte, prot, old_addr, new_addr) (pte) pte 486 include/asm-generic/pgtable.h # define pte_accessible(mm, pte) ((void)(pte), 1) pte 623 include/asm-generic/pgtable.h pte_t *ptep, pte_t pte) pte 629 include/asm-generic/pgtable.h set_pte_at(vma->vm_mm, addr, ptep, pte); pte 660 include/asm-generic/pgtable.h pte_t *ptep, pte_t old_pte, pte_t pte) pte 662 include/asm-generic/pgtable.h __ptep_modify_prot_commit(vma, addr, ptep, pte); pte 733 include/asm-generic/pgtable.h static inline int pte_soft_dirty(pte_t pte) pte 743 include/asm-generic/pgtable.h static inline pte_t pte_mksoft_dirty(pte_t pte) pte 745 include/asm-generic/pgtable.h return pte; pte 753 include/asm-generic/pgtable.h static inline pte_t pte_clear_soft_dirty(pte_t pte) pte 755 include/asm-generic/pgtable.h return pte; pte 763 include/asm-generic/pgtable.h static inline pte_t pte_swp_mksoft_dirty(pte_t pte) pte 765 include/asm-generic/pgtable.h return pte; pte 768 include/asm-generic/pgtable.h static inline int pte_swp_soft_dirty(pte_t pte) pte 773 include/asm-generic/pgtable.h static inline pte_t pte_swp_clear_soft_dirty(pte_t pte) pte 775 include/asm-generic/pgtable.h return pte; pte 1026 include/asm-generic/pgtable.h static inline int pte_protnone(pte_t pte) pte 142 include/linux/hugetlb.h bool is_hugetlb_entry_migration(pte_t pte); pte 537 include/linux/hugetlb.h struct mm_struct *mm, pte_t *pte) pte 540 include/linux/hugetlb.h return pmd_lockptr(mm, (pmd_t *) pte); pte 568 include/linux/hugetlb.h pte_t *ptep, pte_t pte, unsigned long sz) pte 570 include/linux/hugetlb.h set_huge_pte_at(mm, addr, ptep, pte); pte 587 include/linux/hugetlb.h pte_t old_pte, pte_t pte) pte 589 include/linux/hugetlb.h set_huge_pte_at(vma->vm_mm, addr, ptep, pte); pte 723 include/linux/hugetlb.h struct mm_struct *mm, pte_t *pte) pte 737 include/linux/hugetlb.h pte_t *ptep, pte_t pte, unsigned long sz) pte 743 include/linux/hugetlb.h struct mm_struct *mm, pte_t *pte) pte 747 include/linux/hugetlb.h ptl = huge_pte_lockptr(h, mm, pte); pte 606 include/linux/intel-iommu.h static inline void dma_clear_pte(struct dma_pte *pte) pte 608 include/linux/intel-iommu.h pte->val = 0; pte 611 include/linux/intel-iommu.h static inline u64 dma_pte_addr(struct dma_pte *pte) pte 614 include/linux/intel-iommu.h return pte->val & VTD_PAGE_MASK; pte 617 include/linux/intel-iommu.h return __cmpxchg64(&pte->val, 0ULL, 0ULL) & VTD_PAGE_MASK; pte 621 include/linux/intel-iommu.h static inline bool dma_pte_present(struct dma_pte *pte) pte 623 include/linux/intel-iommu.h return (pte->val & 3) != 0; pte 626 include/linux/intel-iommu.h static inline bool dma_pte_superpage(struct dma_pte *pte) pte 628 include/linux/intel-iommu.h return (pte->val & DMA_PTE_LARGE_PAGE); pte 631 include/linux/intel-iommu.h static inline int first_pte_in_page(struct dma_pte *pte) pte 633 include/linux/intel-iommu.h return !((unsigned long)pte & ~VTD_PAGE_MASK); pte 438 include/linux/mm.h pte_t *pte; /* Pointer to pte entry matching pte 842 include/linux/mm.h static inline pte_t maybe_mkwrite(pte_t pte, struct vm_area_struct *vma) pte 845 include/linux/mm.h pte = pte_mkwrite(pte); pte 846 include/linux/mm.h return pte; pte 1452 include/linux/mm.h pte_t pte); pte 1742 include/linux/mm.h static inline int pte_devmap(pte_t pte) pte 1984 include/linux/mm.h #define pte_unmap_unlock(pte, ptl) do { \ pte 1986 include/linux/mm.h pte_unmap(pte); \ pte 2643 include/linux/mm.h typedef int (*pte_fn_t)(pte_t *pte, unsigned long addr, void *data); pte 141 include/linux/mmu_notifier.h pte_t pte); pte 289 include/linux/mmu_notifier.h unsigned long address, pte_t pte); pte 337 include/linux/mmu_notifier.h unsigned long address, pte_t pte) pte 340 include/linux/mmu_notifier.h __mmu_notifier_change_pte(mm, address, pte); pte 576 include/linux/mmu_notifier.h unsigned long address, pte_t pte) pte 33 include/linux/pagewalk.h int (*pte_entry)(pte_t *pte, unsigned long addr, pte 37 include/linux/pagewalk.h int (*hugetlb_entry)(pte_t *pte, unsigned long hmask, pte 112 include/linux/pfn_t.h pte_t pte_mkdevmap(pte_t pte); pte 210 include/linux/rmap.h pte_t *pte; pte 217 include/linux/rmap.h if (pvmw->pte) pte 218 include/linux/rmap.h pte_unmap(pvmw->pte); pte 56 include/linux/swapops.h static inline int is_swap_pte(pte_t pte) pte 58 include/linux/swapops.h return !pte_none(pte) && !pte_present(pte); pte 65 include/linux/swapops.h static inline swp_entry_t pte_to_swp_entry(pte_t pte) pte 69 include/linux/swapops.h if (pte_swp_soft_dirty(pte)) pte 70 include/linux/swapops.h pte = pte_swp_clear_soft_dirty(pte); pte 71 include/linux/swapops.h arch_entry = __pte_to_swp_entry(pte); pte 209 include/linux/swapops.h struct mm_struct *mm, pte_t *pte); pte 234 include/linux/swapops.h struct mm_struct *mm, pte_t *pte) { } pte 13 include/trace/events/thp.h TP_PROTO(unsigned long addr, unsigned long pte), pte 14 include/trace/events/thp.h TP_ARGS(addr, pte), pte 17 include/trace/events/thp.h __field(unsigned long, pte) pte 22 include/trace/events/thp.h __entry->pte = pte; pte 26 include/trace/events/thp.h __entry->addr, __entry->pte) pte 49 include/trace/events/thp.h TP_PROTO(unsigned long addr, unsigned long pte, unsigned long clr, unsigned long set), pte 50 include/trace/events/thp.h TP_ARGS(addr, pte, clr, set), pte 53 include/trace/events/thp.h __field(unsigned long, pte) pte 60 include/trace/events/thp.h __entry->pte = pte; pte 66 include/trace/events/thp.h TP_printk("hugepage update at addr 0x%lx and pte = 0x%lx clr = 0x%lx, set = 0x%lx", __entry->addr, __entry->pte, __entry->clr, __entry->set) pte 70 include/trace/events/thp.h TP_PROTO(unsigned long addr, unsigned long pte), pte 71 include/trace/events/thp.h TP_ARGS(addr, pte), pte 74 include/trace/events/thp.h __field(unsigned long, pte) pte 79 include/trace/events/thp.h __entry->pte = pte; pte 83 include/trace/events/thp.h __entry->addr, __entry->pte) pte 142 include/trace/events/xen.h __entry->pteval = pteval.pte), pte 169 include/trace/events/xen.h __entry->pteval = pteval.pte), pte 291 include/trace/events/xen.h __entry->pteval = pteval.pte), pte 203 kernel/events/uprobes.c flush_cache_page(vma, addr, pte_pfn(*pvmw.pte)); pte 204 kernel/events/uprobes.c ptep_clear_flush_notify(vma, addr, pvmw.pte); pte 206 kernel/events/uprobes.c set_pte_at_notify(mm, addr, pvmw.pte, pte 66 lib/ioremap.c pte_t *pte; pte 70 lib/ioremap.c pte = pte_alloc_kernel(pmd, addr); pte 71 lib/ioremap.c if (!pte) pte 74 lib/ioremap.c BUG_ON(!pte_none(*pte)); pte 75 lib/ioremap.c set_pte_at(&init_mm, addr, pte, pfn_pte(pfn, prot)); pte 77 lib/ioremap.c } while (pte++, addr += PAGE_SIZE, addr != end); pte 2651 mm/filemap.c if (vmf->pte) pte 2652 mm/filemap.c vmf->pte += xas.xa_index - last_pgoff; pte 140 mm/gup.c pte_t *pte, unsigned int flags) pte 147 mm/gup.c pte_t entry = *pte; pte 153 mm/gup.c if (!pte_same(*pte, entry)) { pte 154 mm/gup.c set_pte_at(vma->vm_mm, address, pte, entry); pte 155 mm/gup.c update_mmu_cache(vma, address, pte); pte 167 mm/gup.c static inline bool can_follow_write_pte(pte_t pte, unsigned int flags) pte 169 mm/gup.c return pte_write(pte) || ((flags & FOLL_COW) && pte_dirty(pte)); pte 189 mm/gup.c pte_t *ptep, pte; pte 196 mm/gup.c pte = *ptep; pte 197 mm/gup.c if (!pte_present(pte)) { pte 206 mm/gup.c if (pte_none(pte)) pte 208 mm/gup.c entry = pte_to_swp_entry(pte); pte 215 mm/gup.c if ((flags & FOLL_NUMA) && pte_protnone(pte)) pte 217 mm/gup.c if ((flags & FOLL_WRITE) && !can_follow_write_pte(pte, flags)) { pte 222 mm/gup.c page = vm_normal_page(vma, address, pte); pte 223 mm/gup.c if (!page && pte_devmap(pte) && (flags & FOLL_GET)) { pte 228 mm/gup.c *pgmap = get_dev_pagemap(pte_pfn(pte), *pgmap); pte 230 mm/gup.c page = pte_page(pte); pte 240 mm/gup.c if (is_zero_pfn(pte_pfn(pte))) { pte 241 mm/gup.c page = pte_page(pte); pte 272 mm/gup.c !pte_dirty(pte) && !PageDirty(page)) pte 312 mm/gup.c if (!pte_none(pte)) pte 577 mm/gup.c pte_t *pte; pte 599 mm/gup.c pte = pte_offset_map(pmd, address); pte 600 mm/gup.c if (pte_none(*pte)) pte 605 mm/gup.c *page = vm_normal_page(*vma, address, *pte); pte 607 mm/gup.c if ((gup_flags & FOLL_DUMP) || !is_zero_pfn(pte_pfn(*pte))) pte 609 mm/gup.c *page = pte_page(*pte); pte 618 mm/gup.c pte_unmap(pte); pte 1779 mm/gup.c pte_t pte; pte 1782 mm/gup.c pte.pte_low = ptep->pte_low; pte 1784 mm/gup.c pte.pte_high = ptep->pte_high; pte 1786 mm/gup.c } while (unlikely(pte.pte_low != ptep->pte_low)); pte 1788 mm/gup.c return pte; pte 1835 mm/gup.c pte_t pte = gup_get_pte(ptep); pte 1842 mm/gup.c if (pte_protnone(pte)) pte 1845 mm/gup.c if (!pte_access_permitted(pte, flags & FOLL_WRITE)) pte 1848 mm/gup.c if (pte_devmap(pte)) { pte 1852 mm/gup.c pgmap = get_dev_pagemap(pte_pfn(pte), pgmap); pte 1857 mm/gup.c } else if (pte_special(pte)) pte 1860 mm/gup.c VM_BUG_ON(!pfn_valid(pte_pfn(pte))); pte 1861 mm/gup.c page = pte_page(pte); pte 1867 mm/gup.c if (unlikely(pte_val(pte) != pte_val(*ptep))) { pte 1996 mm/gup.c pte_t pte; pte 2003 mm/gup.c pte = READ_ONCE(*ptep); pte 2005 mm/gup.c if (!pte_access_permitted(pte, flags & FOLL_WRITE)) pte 2009 mm/gup.c VM_BUG_ON(!pfn_valid(pte_pfn(pte))); pte 2012 mm/gup.c head = pte_page(pte); pte 2029 mm/gup.c if (unlikely(pte_val(pte) != pte_val(*ptep))) { pte 445 mm/hmm.c static inline uint64_t pte_to_hmm_pfn_flags(struct hmm_range *range, pte_t pte) pte 447 mm/hmm.c if (pte_none(pte) || !pte_present(pte) || pte_protnone(pte)) pte 449 mm/hmm.c return pte_write(pte) ? range->flags[HMM_PFN_VALID] | pte 462 mm/hmm.c pte_t pte = *ptep; pte 468 mm/hmm.c if (pte_none(pte)) { pte 476 mm/hmm.c if (!pte_present(pte)) { pte 477 mm/hmm.c swp_entry_t entry = pte_to_swp_entry(pte); pte 480 mm/hmm.c cpu_flags = pte_to_hmm_pfn_flags(range, pte); pte 521 mm/hmm.c cpu_flags = pte_to_hmm_pfn_flags(range, pte); pte 529 mm/hmm.c if (pte_devmap(pte)) { pte 530 mm/hmm.c hmm_vma_walk->pgmap = get_dev_pagemap(pte_pfn(pte), pte 534 mm/hmm.c } else if (IS_ENABLED(CONFIG_ARCH_HAS_PTE_SPECIAL) && pte_special(pte)) { pte 539 mm/hmm.c *pfn = hmm_device_entry_from_pfn(range, pte_pfn(pte)) | cpu_flags; pte 726 mm/hmm.c static int hmm_vma_walk_hugetlb_entry(pte_t *pte, unsigned long hmask, pte 740 mm/hmm.c ptl = huge_pte_lock(hstate_vma(vma), walk->mm, pte); pte 741 mm/hmm.c entry = huge_ptep_get(pte); pte 1278 mm/huge_memory.c vmf->pte = pte_offset_map(&_pmd, haddr); pte 1279 mm/huge_memory.c VM_BUG_ON(!pte_none(*vmf->pte)); pte 1280 mm/huge_memory.c set_pte_at(vma->vm_mm, haddr, vmf->pte, entry); pte 1281 mm/huge_memory.c pte_unmap(vmf->pte); pte 2127 mm/huge_memory.c pte_t *pte, entry; pte 2130 mm/huge_memory.c pte = pte_offset_map(&_pmd, haddr); pte 2131 mm/huge_memory.c VM_BUG_ON(!pte_none(*pte)); pte 2132 mm/huge_memory.c set_pte_at(mm, haddr, pte, entry); pte 2133 mm/huge_memory.c pte_unmap(pte); pte 2240 mm/huge_memory.c pte_t entry, *pte; pte 2262 mm/huge_memory.c pte = pte_offset_map(&_pmd, addr); pte 2263 mm/huge_memory.c BUG_ON(!pte_none(*pte)); pte 2264 mm/huge_memory.c set_pte_at(mm, addr, pte, entry); pte 2266 mm/huge_memory.c pte_unmap(pte); pte 3030 mm/huge_memory.c if (!(pvmw->pmd && !pvmw->pte)) pte 3055 mm/huge_memory.c if (!(pvmw->pmd && !pvmw->pte)) pte 3393 mm/hugetlb.c bool is_hugetlb_entry_migration(pte_t pte) pte 3397 mm/hugetlb.c if (huge_pte_none(pte) || pte_present(pte)) pte 3399 mm/hugetlb.c swp = pte_to_swp_entry(pte); pte 3406 mm/hugetlb.c static int is_hugetlb_entry_hwpoisoned(pte_t pte) pte 3410 mm/hugetlb.c if (huge_pte_none(pte) || pte_present(pte)) pte 3412 mm/hugetlb.c swp = pte_to_swp_entry(pte); pte 3526 mm/hugetlb.c pte_t pte; pte 3567 mm/hugetlb.c pte = huge_ptep_get(ptep); pte 3568 mm/hugetlb.c if (huge_pte_none(pte)) { pte 3577 mm/hugetlb.c if (unlikely(!pte_present(pte))) { pte 3583 mm/hugetlb.c page = pte_page(pte); pte 3602 mm/hugetlb.c pte = huge_ptep_get_and_clear(mm, address, ptep); pte 3604 mm/hugetlb.c if (huge_pte_dirty(pte)) pte 3731 mm/hugetlb.c pte_t pte; pte 3739 mm/hugetlb.c pte = huge_ptep_get(ptep); pte 3740 mm/hugetlb.c old_page = pte_page(pte); pte 3783 mm/hugetlb.c BUG_ON(huge_pte_none(pte)); pte 3785 mm/hugetlb.c BUG_ON(huge_pte_none(pte)); pte 3789 mm/hugetlb.c pte_same(huge_ptep_get(ptep), pte))) pte 3825 mm/hugetlb.c if (likely(ptep && pte_same(huge_ptep_get(ptep), pte))) { pte 4399 mm/hugetlb.c pte_t *pte; pte 4420 mm/hugetlb.c pte = huge_pte_offset(mm, vaddr & huge_page_mask(h), pte 4422 mm/hugetlb.c if (pte) pte 4423 mm/hugetlb.c ptl = huge_pte_lock(h, mm, pte); pte 4424 mm/hugetlb.c absent = !pte || huge_pte_none(huge_ptep_get(pte)); pte 4435 mm/hugetlb.c if (pte) pte 4451 mm/hugetlb.c if (absent || is_swap_pte(huge_ptep_get(pte)) || pte 4453 mm/hugetlb.c !huge_pte_write(huge_ptep_get(pte)))) { pte 4457 mm/hugetlb.c if (pte) pte 4497 mm/hugetlb.c page = pte_page(huge_ptep_get(pte)); pte 4559 mm/hugetlb.c pte_t pte; pte 4591 mm/hugetlb.c pte = huge_ptep_get(ptep); pte 4592 mm/hugetlb.c if (unlikely(is_hugetlb_entry_hwpoisoned(pte))) { pte 4596 mm/hugetlb.c if (unlikely(is_hugetlb_entry_migration(pte))) { pte 4597 mm/hugetlb.c swp_entry_t entry = pte_to_swp_entry(pte); pte 4611 mm/hugetlb.c if (!huge_pte_none(pte)) { pte 4615 mm/hugetlb.c pte = pte_mkhuge(huge_pte_modify(old_pte, newprot)); pte 4616 mm/hugetlb.c pte = arch_make_huge_pte(pte, vma, NULL, 0); pte 4617 mm/hugetlb.c huge_ptep_modify_prot_commit(vma, address, ptep, old_pte, pte); pte 4888 mm/hugetlb.c pte_t *pte; pte 4923 mm/hugetlb.c pte = (pte_t *)pmd_alloc(mm, pud, addr); pte 4925 mm/hugetlb.c return pte; pte 4982 mm/hugetlb.c pte_t *pte = NULL; pte 4991 mm/hugetlb.c pte = (pte_t *)pud; pte 4995 mm/hugetlb.c pte = huge_pmd_share(mm, addr, pud); pte 4997 mm/hugetlb.c pte = (pte_t *)pmd_alloc(mm, pud, addr); pte 5000 mm/hugetlb.c BUG_ON(pte && pte_present(*pte) && !pte_huge(*pte)); pte 5002 mm/hugetlb.c return pte; pte 5075 mm/hugetlb.c pte_t pte; pte 5085 mm/hugetlb.c pte = huge_ptep_get((pte_t *)pmd); pte 5086 mm/hugetlb.c if (pte_present(pte)) { pte 5091 mm/hugetlb.c if (is_hugetlb_entry_migration(pte)) { pte 645 mm/kasan/common.c pte_t *pte; pte 669 mm/kasan/common.c pte = pte_offset_kernel(pmd, addr); pte 670 mm/kasan/common.c return !pte_none(*pte); pte 79 mm/kasan/init.c static inline bool kasan_early_shadow_page_entry(pte_t pte) pte 81 mm/kasan/init.c return pte_page(pte) == virt_to_page(lm_alias(kasan_early_shadow_page)); pte 99 mm/kasan/init.c pte_t *pte = pte_offset_kernel(pmd, addr); pte 107 mm/kasan/init.c set_pte_at(&init_mm, addr, pte, zero_pte); pte 109 mm/kasan/init.c pte = pte_offset_kernel(pmd, addr); pte 299 mm/kasan/init.c pte_t *pte; pte 303 mm/kasan/init.c pte = pte_start + i; pte 304 mm/kasan/init.c if (!pte_none(*pte)) pte 357 mm/kasan/init.c static void kasan_remove_pte_table(pte_t *pte, unsigned long addr, pte 362 mm/kasan/init.c for (; addr < end; addr = next, pte++) { pte 367 mm/kasan/init.c if (!pte_present(*pte)) pte 370 mm/kasan/init.c if (WARN_ON(!kasan_early_shadow_page_entry(*pte))) pte 372 mm/kasan/init.c pte_clear(&init_mm, addr, pte); pte 382 mm/kasan/init.c pte_t *pte; pte 395 mm/kasan/init.c pte = pte_offset_kernel(pmd, addr); pte 396 mm/kasan/init.c kasan_remove_pte_table(pte, addr, next); pte 523 mm/khugepaged.c static void release_pte_pages(pte_t *pte, pte_t *_pte) pte 525 mm/khugepaged.c while (--_pte >= pte) { pte 534 mm/khugepaged.c pte_t *pte) pte 541 mm/khugepaged.c for (_pte = pte; _pte < pte+HPAGE_PMD_NR; pte 640 mm/khugepaged.c release_pte_pages(pte, _pte); pte 646 mm/khugepaged.c static void __collapse_huge_page_copy(pte_t *pte, struct page *page, pte 652 mm/khugepaged.c for (_pte = pte; _pte < pte + HPAGE_PMD_NR; pte 910 mm/khugepaged.c vmf.pte = pte_offset_map(pmd, address); pte 912 mm/khugepaged.c vmf.pte++, vmf.address += PAGE_SIZE) { pte 913 mm/khugepaged.c vmf.orig_pte = *vmf.pte; pte 938 mm/khugepaged.c vmf.pte = pte_offset_map(pmd, vmf.address); pte 940 mm/khugepaged.c vmf.pte--; pte 941 mm/khugepaged.c pte_unmap(vmf.pte); pte 952 mm/khugepaged.c pte_t *pte; pte 1035 mm/khugepaged.c pte = pte_offset_map(pmd, address); pte 1050 mm/khugepaged.c isolated = __collapse_huge_page_isolate(vma, address, pte); pte 1054 mm/khugepaged.c pte_unmap(pte); pte 1075 mm/khugepaged.c __collapse_huge_page_copy(pte, new_page, vma, address, pte_ptl); pte 1076 mm/khugepaged.c pte_unmap(pte); pte 1121 mm/khugepaged.c pte_t *pte, *_pte; pte 1138 mm/khugepaged.c pte = pte_offset_map_lock(mm, pmd, address, &ptl); pte 1139 mm/khugepaged.c for (_address = address, _pte = pte; _pte < pte+HPAGE_PMD_NR; pte 1228 mm/khugepaged.c pte_unmap_unlock(pte, ptl); pte 1295 mm/khugepaged.c pte_t *start_pte, *pte; pte 1321 mm/khugepaged.c for (i = 0, addr = haddr, pte = start_pte; pte 1322 mm/khugepaged.c i < HPAGE_PMD_NR; i++, addr += PAGE_SIZE, pte++) { pte 1326 mm/khugepaged.c if (pte_none(*pte)) pte 1330 mm/khugepaged.c if (!pte_present(*pte)) pte 1333 mm/khugepaged.c page = vm_normal_page(vma, addr, *pte); pte 1364 mm/khugepaged.c for (i = 0, addr = haddr, pte = start_pte; pte 1365 mm/khugepaged.c i < HPAGE_PMD_NR; i++, addr += PAGE_SIZE, pte++) { pte 1368 mm/khugepaged.c if (pte_none(*pte)) pte 1370 mm/khugepaged.c page = vm_normal_page(vma, addr, *pte); pte 1057 mm/ksm.c if (WARN_ONCE(!pvmw.pte, "Unexpected PMD mapping?")) pte 1060 mm/ksm.c if (pte_write(*pvmw.pte) || pte_dirty(*pvmw.pte) || pte 1061 mm/ksm.c (pte_protnone(*pvmw.pte) && pte_savedwrite(*pvmw.pte)) || pte 1081 mm/ksm.c entry = ptep_clear_flush(vma, pvmw.address, pvmw.pte); pte 1087 mm/ksm.c set_pte_at(mm, pvmw.address, pvmw.pte, entry); pte 1097 mm/ksm.c set_pte_at_notify(mm, pvmw.address, pvmw.pte, entry); pte 1099 mm/ksm.c *orig_pte = *pvmw.pte; pte 194 mm/madvise.c pte_t pte; pte 200 mm/madvise.c pte = *(orig_pte + ((index - start) / PAGE_SIZE)); pte 203 mm/madvise.c if (pte_present(pte) || pte_none(pte)) pte 205 mm/madvise.c entry = pte_to_swp_entry(pte); pte 309 mm/madvise.c pte_t *orig_pte, *pte, ptent; pte 388 mm/madvise.c orig_pte = pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl); pte 391 mm/madvise.c for (; addr < end; pte++, addr += PAGE_SIZE) { pte 392 mm/madvise.c ptent = *pte; pte 425 mm/madvise.c pte = pte_offset_map_lock(mm, pmd, addr, &ptl); pte 426 mm/madvise.c pte--; pte 438 mm/madvise.c ptent = ptep_get_and_clear_full(mm, addr, pte, pte 441 mm/madvise.c set_pte_at(mm, addr, pte, ptent); pte 442 mm/madvise.c tlb_remove_tlb_entry(tlb, pte, addr); pte 570 mm/madvise.c pte_t *orig_pte, *pte, ptent; pte 584 mm/madvise.c orig_pte = pte = pte_offset_map_lock(mm, pmd, addr, &ptl); pte 587 mm/madvise.c for (; addr != end; pte++, addr += PAGE_SIZE) { pte 588 mm/madvise.c ptent = *pte; pte 605 mm/madvise.c pte_clear_not_present_full(mm, addr, pte, tlb->fullmm); pte 635 mm/madvise.c pte = pte_offset_map_lock(mm, pmd, addr, &ptl); pte 636 mm/madvise.c pte--; pte 671 mm/madvise.c ptent = ptep_get_and_clear_full(mm, addr, pte, pte 676 mm/madvise.c set_pte_at(mm, addr, pte, ptent); pte 677 mm/madvise.c tlb_remove_tlb_entry(tlb, pte, addr); pte 5682 mm/memcontrol.c pte_t *pte; pte 5700 mm/memcontrol.c pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl); pte 5701 mm/memcontrol.c for (; addr != end; pte++, addr += PAGE_SIZE) pte 5702 mm/memcontrol.c if (get_mctgt_type(vma, addr, *pte, NULL)) pte 5704 mm/memcontrol.c pte_unmap_unlock(pte - 1, ptl); pte 5882 mm/memcontrol.c pte_t *pte; pte 5922 mm/memcontrol.c pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl); pte 5924 mm/memcontrol.c pte_t ptent = *(pte++); pte 5970 mm/memcontrol.c pte_unmap_unlock(pte - 1, ptl); pte 272 mm/memory-failure.c pte_t *pte; pte 290 mm/memory-failure.c pte = pte_offset_map(pmd, address); pte 291 mm/memory-failure.c if (!pte_present(*pte)) pte 293 mm/memory-failure.c if (pte_devmap(*pte)) pte 482 mm/memory.c pte_t pte, struct page *page) pte 518 mm/memory.c (long long)pte_val(pte), (long long)pmd_val(*pmd)); pte 575 mm/memory.c pte_t pte) pte 577 mm/memory.c unsigned long pfn = pte_pfn(pte); pte 580 mm/memory.c if (likely(!pte_special(pte))) pte 588 mm/memory.c if (pte_devmap(pte)) pte 591 mm/memory.c print_bad_pte(vma, addr, pte, NULL); pte 617 mm/memory.c print_bad_pte(vma, addr, pte, NULL); pte 683 mm/memory.c pte_t pte = *src_pte; pte 687 mm/memory.c if (unlikely(!pte_present(pte))) { pte 688 mm/memory.c swp_entry_t entry = pte_to_swp_entry(pte); pte 715 mm/memory.c pte = swp_entry_to_pte(entry); pte 717 mm/memory.c pte = pte_swp_mksoft_dirty(pte); pte 718 mm/memory.c set_pte_at(src_mm, addr, src_pte, pte); pte 746 mm/memory.c pte = swp_entry_to_pte(entry); pte 747 mm/memory.c set_pte_at(src_mm, addr, src_pte, pte); pte 757 mm/memory.c if (is_cow_mapping(vm_flags) && pte_write(pte)) { pte 759 mm/memory.c pte = pte_wrprotect(pte); pte 767 mm/memory.c pte = pte_mkclean(pte); pte 768 mm/memory.c pte = pte_mkold(pte); pte 770 mm/memory.c page = vm_normal_page(vma, addr, pte); pte 775 mm/memory.c } else if (pte_devmap(pte)) { pte 776 mm/memory.c page = pte_page(pte); pte 780 mm/memory.c set_pte_at(dst_mm, addr, dst_pte, pte); pte 1014 mm/memory.c pte_t *pte; pte 1021 mm/memory.c pte = start_pte; pte 1025 mm/memory.c pte_t ptent = *pte; pte 1046 mm/memory.c ptent = ptep_get_and_clear_full(mm, addr, pte, pte 1048 mm/memory.c tlb_remove_tlb_entry(tlb, pte, addr); pte 1088 mm/memory.c pte_clear_not_present_full(mm, addr, pte, tlb->fullmm); pte 1109 mm/memory.c pte_clear_not_present_full(mm, addr, pte, tlb->fullmm); pte 1110 mm/memory.c } while (pte++, addr += PAGE_SIZE, addr != end); pte 1427 mm/memory.c pte_t *pte; pte 1435 mm/memory.c pte = get_locked_pte(mm, addr, &ptl); pte 1436 mm/memory.c if (!pte) pte 1439 mm/memory.c if (!pte_none(*pte)) pte 1446 mm/memory.c set_pte_at(mm, addr, pte, mk_pte(page, prot)); pte 1450 mm/memory.c pte_unmap_unlock(pte, ptl); pte 1585 mm/memory.c pte_t *pte, entry; pte 1588 mm/memory.c pte = get_locked_pte(mm, addr, &ptl); pte 1589 mm/memory.c if (!pte) pte 1591 mm/memory.c if (!pte_none(*pte)) { pte 1603 mm/memory.c if (pte_pfn(*pte) != pfn_t_to_pfn(pfn)) { pte 1604 mm/memory.c WARN_ON_ONCE(!is_zero_pfn(pte_pfn(*pte))); pte 1607 mm/memory.c entry = pte_mkyoung(*pte); pte 1609 mm/memory.c if (ptep_set_access_flags(vma, addr, pte, entry, 1)) pte 1610 mm/memory.c update_mmu_cache(vma, addr, pte); pte 1626 mm/memory.c set_pte_at(mm, addr, pte, entry); pte 1627 mm/memory.c update_mmu_cache(vma, addr, pte); /* XXX: why not for insert_page? */ pte 1630 mm/memory.c pte_unmap_unlock(pte, ptl); pte 1795 mm/memory.c pte_t *pte; pte 1799 mm/memory.c pte = pte_alloc_map_lock(mm, pmd, addr, &ptl); pte 1800 mm/memory.c if (!pte) pte 1804 mm/memory.c BUG_ON(!pte_none(*pte)); pte 1809 mm/memory.c set_pte_at(mm, addr, pte, pte_mkspecial(pfn_pte(pfn, prot))); pte 1811 mm/memory.c } while (pte++, addr += PAGE_SIZE, addr != end); pte 1813 mm/memory.c pte_unmap_unlock(pte - 1, ptl); pte 2008 mm/memory.c pte_t *pte; pte 2012 mm/memory.c pte = (mm == &init_mm) ? pte 2015 mm/memory.c if (!pte) pte 2023 mm/memory.c err = fn(pte++, addr, data); pte 2031 mm/memory.c pte_unmap_unlock(pte-1, ptl); pte 2300 mm/memory.c if (ptep_set_access_flags(vma, vmf->address, vmf->pte, entry, 1)) pte 2301 mm/memory.c update_mmu_cache(vma, vmf->address, vmf->pte); pte 2302 mm/memory.c pte_unmap_unlock(vmf->pte, vmf->ptl); pte 2361 mm/memory.c vmf->pte = pte_offset_map_lock(mm, vmf->pmd, vmf->address, &vmf->ptl); pte 2362 mm/memory.c if (likely(pte_same(*vmf->pte, vmf->orig_pte))) { pte 2381 mm/memory.c ptep_clear_flush_notify(vma, vmf->address, vmf->pte); pte 2390 mm/memory.c set_pte_at_notify(mm, vmf->address, vmf->pte, entry); pte 2391 mm/memory.c update_mmu_cache(vma, vmf->address, vmf->pte); pte 2428 mm/memory.c pte_unmap_unlock(vmf->pte, vmf->ptl); pte 2475 mm/memory.c vmf->pte = pte_offset_map_lock(vmf->vma->vm_mm, vmf->pmd, vmf->address, pte 2481 mm/memory.c if (!pte_same(*vmf->pte, vmf->orig_pte)) { pte 2482 mm/memory.c pte_unmap_unlock(vmf->pte, vmf->ptl); pte 2500 mm/memory.c pte_unmap_unlock(vmf->pte, vmf->ptl); pte 2522 mm/memory.c pte_unmap_unlock(vmf->pte, vmf->ptl); pte 2581 mm/memory.c pte_unmap_unlock(vmf->pte, vmf->ptl); pte 2596 mm/memory.c pte_unmap_unlock(vmf->pte, vmf->ptl); pte 2598 mm/memory.c vmf->pte = pte_offset_map_lock(vma->vm_mm, vmf->pmd, pte 2600 mm/memory.c if (!pte_same(*vmf->pte, vmf->orig_pte)) { pte 2602 mm/memory.c pte_unmap_unlock(vmf->pte, vmf->ptl); pte 2643 mm/memory.c pte_unmap_unlock(vmf->pte, vmf->ptl); pte 2757 mm/memory.c pte_t pte; pte 2762 mm/memory.c if (!pte_unmap_same(vma->vm_mm, vmf->pmd, vmf->pte, vmf->orig_pte)) pte 2813 mm/memory.c vmf->pte = pte_offset_map_lock(vma->vm_mm, vmf->pmd, pte 2815 mm/memory.c if (likely(pte_same(*vmf->pte, vmf->orig_pte))) pte 2869 mm/memory.c vmf->pte = pte_offset_map_lock(vma->vm_mm, vmf->pmd, vmf->address, pte 2871 mm/memory.c if (unlikely(!pte_same(*vmf->pte, vmf->orig_pte))) pte 2891 mm/memory.c pte = mk_pte(page, vma->vm_page_prot); pte 2893 mm/memory.c pte = maybe_mkwrite(pte_mkdirty(pte), vma); pte 2900 mm/memory.c pte = pte_mksoft_dirty(pte); pte 2901 mm/memory.c set_pte_at(vma->vm_mm, vmf->address, vmf->pte, pte); pte 2902 mm/memory.c arch_do_swap_page(vma->vm_mm, vma, vmf->address, pte, vmf->orig_pte); pte 2903 mm/memory.c vmf->orig_pte = pte; pte 2942 mm/memory.c update_mmu_cache(vma, vmf->address, vmf->pte); pte 2944 mm/memory.c pte_unmap_unlock(vmf->pte, vmf->ptl); pte 2949 mm/memory.c pte_unmap_unlock(vmf->pte, vmf->ptl); pte 3000 mm/memory.c vmf->pte = pte_offset_map_lock(vma->vm_mm, vmf->pmd, pte 3002 mm/memory.c if (!pte_none(*vmf->pte)) pte 3009 mm/memory.c pte_unmap_unlock(vmf->pte, vmf->ptl); pte 3037 mm/memory.c vmf->pte = pte_offset_map_lock(vma->vm_mm, vmf->pmd, vmf->address, pte 3039 mm/memory.c if (!pte_none(*vmf->pte)) pte 3048 mm/memory.c pte_unmap_unlock(vmf->pte, vmf->ptl); pte 3059 mm/memory.c set_pte_at(vma->vm_mm, vmf->address, vmf->pte, entry); pte 3062 mm/memory.c update_mmu_cache(vma, vmf->address, vmf->pte); pte 3064 mm/memory.c pte_unmap_unlock(vmf->pte, vmf->ptl); pte 3184 mm/memory.c vmf->pte = pte_offset_map_lock(vma->vm_mm, vmf->pmd, vmf->address, pte 3301 mm/memory.c if (!vmf->pte) { pte 3308 mm/memory.c if (unlikely(!pte_none(*vmf->pte))) pte 3325 mm/memory.c set_pte_at(vma->vm_mm, vmf->address, vmf->pte, entry); pte 3328 mm/memory.c update_mmu_cache(vma, vmf->address, vmf->pte); pte 3369 mm/memory.c if (vmf->pte) pte 3370 mm/memory.c pte_unmap_unlock(vmf->pte, vmf->ptl); pte 3475 mm/memory.c if (!vmf->pte) pte 3479 mm/memory.c vmf->pte -= (vmf->address >> PAGE_SHIFT) - (address >> PAGE_SHIFT); pte 3480 mm/memory.c if (!pte_none(*vmf->pte)) pte 3482 mm/memory.c pte_unmap_unlock(vmf->pte, vmf->ptl); pte 3485 mm/memory.c vmf->pte = NULL; pte 3615 mm/memory.c vmf->pte = pte_offset_map_lock(vmf->vma->vm_mm, pte 3626 mm/memory.c if (unlikely(pte_none(*vmf->pte))) pte 3631 mm/memory.c pte_unmap_unlock(vmf->pte, vmf->ptl); pte 3671 mm/memory.c pte_t pte, old_pte; pte 3682 mm/memory.c if (unlikely(!pte_same(*vmf->pte, vmf->orig_pte))) { pte 3683 mm/memory.c pte_unmap_unlock(vmf->pte, vmf->ptl); pte 3691 mm/memory.c old_pte = ptep_modify_prot_start(vma, vmf->address, vmf->pte); pte 3692 mm/memory.c pte = pte_modify(old_pte, vma->vm_page_prot); pte 3693 mm/memory.c pte = pte_mkyoung(pte); pte 3695 mm/memory.c pte = pte_mkwrite(pte); pte 3696 mm/memory.c ptep_modify_prot_commit(vma, vmf->address, vmf->pte, old_pte, pte); pte 3697 mm/memory.c update_mmu_cache(vma, vmf->address, vmf->pte); pte 3699 mm/memory.c page = vm_normal_page(vma, vmf->address, pte); pte 3701 mm/memory.c pte_unmap_unlock(vmf->pte, vmf->ptl); pte 3707 mm/memory.c pte_unmap_unlock(vmf->pte, vmf->ptl); pte 3719 mm/memory.c if (!pte_write(pte)) pte 3733 mm/memory.c pte_unmap_unlock(vmf->pte, vmf->ptl); pte 3832 mm/memory.c vmf->pte = NULL; pte 3843 mm/memory.c vmf->pte = pte_offset_map(vmf->pmd, vmf->address); pte 3844 mm/memory.c vmf->orig_pte = *vmf->pte; pte 3856 mm/memory.c pte_unmap(vmf->pte); pte 3857 mm/memory.c vmf->pte = NULL; pte 3861 mm/memory.c if (!vmf->pte) { pte 3877 mm/memory.c if (unlikely(!pte_same(*vmf->pte, entry))) pte 3885 mm/memory.c if (ptep_set_access_flags(vmf->vma, vmf->address, vmf->pte, entry, pte 3887 mm/memory.c update_mmu_cache(vmf->vma, vmf->address, vmf->pte); pte 3899 mm/memory.c pte_unmap_unlock(vmf->pte, vmf->ptl); pte 4260 mm/memory.c pte_t *ptep, pte; pte 4268 mm/memory.c pte = *ptep; pte 4270 mm/memory.c if ((flags & FOLL_WRITE) && !pte_write(pte)) pte 4273 mm/memory.c *prot = pgprot_val(pte_pgprot(pte)); pte 4274 mm/memory.c *phys = (resource_size_t)pte_pfn(pte) << PAGE_SHIFT; pte 499 mm/mempolicy.c pte_t *pte; pte 513 mm/mempolicy.c pte = pte_offset_map_lock(walk->mm, pmd, addr, &ptl); pte 514 mm/mempolicy.c for (; addr != end; pte++, addr += PAGE_SIZE) { pte 515 mm/mempolicy.c if (!pte_present(*pte)) pte 517 mm/mempolicy.c page = vm_normal_page(vma, addr, *pte); pte 545 mm/mempolicy.c pte_unmap_unlock(pte - 1, ptl); pte 554 mm/mempolicy.c static int queue_pages_hugetlb(pte_t *pte, unsigned long hmask, pte 565 mm/mempolicy.c ptl = huge_pte_lock(hstate_vma(walk->vma), walk->mm, pte); pte 566 mm/mempolicy.c entry = huge_ptep_get(pte); pte 214 mm/migrate.c pte_t pte; pte 227 mm/migrate.c if (!pvmw.pte) { pte 235 mm/migrate.c pte = pte_mkold(mk_pte(new, READ_ONCE(vma->vm_page_prot))); pte 236 mm/migrate.c if (pte_swp_soft_dirty(*pvmw.pte)) pte 237 mm/migrate.c pte = pte_mksoft_dirty(pte); pte 242 mm/migrate.c entry = pte_to_swp_entry(*pvmw.pte); pte 244 mm/migrate.c pte = maybe_mkwrite(pte, vma); pte 248 mm/migrate.c entry = make_device_private_entry(new, pte_write(pte)); pte 249 mm/migrate.c pte = swp_entry_to_pte(entry); pte 255 mm/migrate.c pte = pte_mkhuge(pte); pte 256 mm/migrate.c pte = arch_make_huge_pte(pte, vma, new, 0); pte 257 mm/migrate.c set_huge_pte_at(vma->vm_mm, pvmw.address, pvmw.pte, pte); pte 265 mm/migrate.c set_pte_at(vma->vm_mm, pvmw.address, pvmw.pte, pte); pte 279 mm/migrate.c update_mmu_cache(vma, pvmw.address, pvmw.pte); pte 310 mm/migrate.c pte_t pte; pte 315 mm/migrate.c pte = *ptep; pte 316 mm/migrate.c if (!is_swap_pte(pte)) pte 319 mm/migrate.c entry = pte_to_swp_entry(pte); pte 348 mm/migrate.c struct mm_struct *mm, pte_t *pte) pte 350 mm/migrate.c spinlock_t *ptl = huge_pte_lockptr(hstate_vma(vma), mm, pte); pte 351 mm/migrate.c __migration_entry_wait(mm, pte, ptl); pte 2250 mm/migrate.c pte_t pte; pte 2252 mm/migrate.c pte = *ptep; pte 2254 mm/migrate.c if (pte_none(pte)) { pte 2260 mm/migrate.c if (!pte_present(pte)) { pte 2268 mm/migrate.c entry = pte_to_swp_entry(pte); pte 2278 mm/migrate.c pfn = pte_pfn(pte); pte 2284 mm/migrate.c page = vm_normal_page(migrate->vma, addr, pte); pte 2286 mm/migrate.c mpfn |= pte_write(pte) ? MIGRATE_PFN_WRITE : 0; pte 2322 mm/migrate.c if (pte_soft_dirty(pte)) pte 2334 mm/migrate.c if (pte_present(pte)) pte 24 mm/mincore.c static int mincore_hugetlb(pte_t *pte, unsigned long hmask, unsigned long addr, pte 35 mm/mincore.c present = pte && !huge_pte_none(huge_ptep_get(pte)); pte 145 mm/mincore.c pte_t pte = *ptep; pte 147 mm/mincore.c if (pte_none(pte)) pte 150 mm/mincore.c else if (pte_present(pte)) pte 153 mm/mincore.c swp_entry_t entry = pte_to_swp_entry(pte); pte 378 mm/mlock.c pte_t *pte; pte 386 mm/mlock.c pte = get_locked_pte(vma->vm_mm, start, &ptl); pte 397 mm/mlock.c pte++; pte 398 mm/mlock.c if (pte_present(*pte)) pte 399 mm/mlock.c page = vm_normal_page(vma, start, *pte); pte 423 mm/mlock.c pte_unmap_unlock(pte, ptl); pte 149 mm/mmu_notifier.c pte_t pte) pte 157 mm/mmu_notifier.c mn->ops->change_pte(mn, mm, address, pte); pte 42 mm/mprotect.c pte_t *pte, oldpte; pte 61 mm/mprotect.c pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl); pte 71 mm/mprotect.c oldpte = *pte; pte 112 mm/mprotect.c oldpte = ptep_modify_prot_start(vma, addr, pte); pte 123 mm/mprotect.c ptep_modify_prot_commit(vma, addr, pte, oldpte, ptent); pte 138 mm/mprotect.c set_pte_at(vma->vm_mm, addr, pte, newpte); pte 152 mm/mprotect.c set_pte_at(vma->vm_mm, addr, pte, newpte); pte 157 mm/mprotect.c } while (pte++, addr += PAGE_SIZE, addr != end); pte 159 mm/mprotect.c pte_unmap_unlock(pte - 1, ptl); pte 345 mm/mprotect.c static int prot_none_pte_entry(pte_t *pte, unsigned long addr, pte 348 mm/mprotect.c return pfn_modify_allowed(pte_pfn(*pte), *(pgprot_t *)(walk->private)) ? pte 352 mm/mprotect.c static int prot_none_hugetlb_entry(pte_t *pte, unsigned long hmask, pte 356 mm/mprotect.c return pfn_modify_allowed(pte_pfn(*pte), *(pgprot_t *)(walk->private)) ? pte 100 mm/mremap.c static pte_t move_soft_dirty_pte(pte_t pte) pte 107 mm/mremap.c if (pte_present(pte)) pte 108 mm/mremap.c pte = pte_mksoft_dirty(pte); pte 109 mm/mremap.c else if (is_swap_pte(pte)) pte 110 mm/mremap.c pte = pte_swp_mksoft_dirty(pte); pte 112 mm/mremap.c return pte; pte 121 mm/mremap.c pte_t *old_pte, *new_pte, pte; pte 164 mm/mremap.c pte = ptep_get_and_clear(mm, old_addr, old_pte); pte 176 mm/mremap.c if (pte_present(pte)) pte 178 mm/mremap.c pte = move_pte(pte, new_vma->vm_page_prot, old_addr, new_addr); pte 179 mm/mremap.c pte = move_soft_dirty_pte(pte); pte 180 mm/mremap.c set_pte_at(mm, new_addr, new_pte, pte); pte 67 mm/page_idle.c if (pvmw.pte) { pte 72 mm/page_idle.c if (ptep_clear_young_notify(vma, addr, pvmw.pte)) pte 18 mm/page_vma_mapped.c pvmw->pte = pte_offset_map(pvmw->pmd, pvmw->address); pte 21 mm/page_vma_mapped.c if (!is_swap_pte(*pvmw->pte)) pte 39 mm/page_vma_mapped.c if (is_swap_pte(*pvmw->pte)) { pte 43 mm/page_vma_mapped.c entry = pte_to_swp_entry(*pvmw->pte); pte 46 mm/page_vma_mapped.c } else if (!pte_present(*pvmw->pte)) pte 87 mm/page_vma_mapped.c if (!is_swap_pte(*pvmw->pte)) pte 89 mm/page_vma_mapped.c entry = pte_to_swp_entry(*pvmw->pte); pte 95 mm/page_vma_mapped.c } else if (is_swap_pte(*pvmw->pte)) { pte 99 mm/page_vma_mapped.c entry = pte_to_swp_entry(*pvmw->pte); pte 105 mm/page_vma_mapped.c if (!pte_present(*pvmw->pte)) pte 108 mm/page_vma_mapped.c pfn = pte_pfn(*pvmw->pte); pte 148 mm/page_vma_mapped.c if (pvmw->pmd && !pvmw->pte) pte 151 mm/page_vma_mapped.c if (pvmw->pte) pte 156 mm/page_vma_mapped.c pvmw->pte = huge_pte_offset(mm, pvmw->address, page_size(page)); pte 157 mm/page_vma_mapped.c if (!pvmw->pte) pte 160 mm/page_vma_mapped.c pvmw->ptl = huge_pte_lockptr(page_hstate(page), mm, pvmw->pte); pte 230 mm/page_vma_mapped.c pte_unmap(pvmw->pte); pte 237 mm/page_vma_mapped.c pvmw->pte++; pte 239 mm/page_vma_mapped.c } while (pte_none(*pvmw->pte)); pte 10 mm/pagewalk.c pte_t *pte; pte 14 mm/pagewalk.c pte = pte_offset_map(pmd, addr); pte 16 mm/pagewalk.c err = ops->pte_entry(pte, addr, addr + PAGE_SIZE, walk); pte 22 mm/pagewalk.c pte++; pte 25 mm/pagewalk.c pte_unmap(pte); pte 190 mm/pagewalk.c pte_t *pte; pte 196 mm/pagewalk.c pte = huge_pte_offset(walk->mm, addr & hmask, sz); pte 198 mm/pagewalk.c if (pte) pte 199 mm/pagewalk.c err = ops->hugetlb_entry(pte, hmask, addr, next, walk); pte 85 mm/pgtable-generic.c pte_t pte; pte 86 mm/pgtable-generic.c pte = ptep_get_and_clear(mm, address, ptep); pte 87 mm/pgtable-generic.c if (pte_accessible(mm, pte)) pte 89 mm/pgtable-generic.c return pte; pte 774 mm/rmap.c if (pvmw.pte) { pte 776 mm/rmap.c pvmw.pte)) { pte 909 mm/rmap.c if (pvmw.pte) { pte 911 mm/rmap.c pte_t *pte = pvmw.pte; pte 913 mm/rmap.c if (!pte_dirty(*pte) && !pte_write(*pte)) pte 916 mm/rmap.c flush_cache_page(vma, address, pte_pfn(*pte)); pte 917 mm/rmap.c entry = ptep_clear_flush(vma, address, pte); pte 920 mm/rmap.c set_pte_at(vma->vm_mm, address, pte, entry); pte 1393 mm/rmap.c if (!pvmw.pte && (flags & TTU_MIGRATION)) { pte 1425 mm/rmap.c VM_BUG_ON_PAGE(!pvmw.pte, page); pte 1427 mm/rmap.c subpage = page - page_to_pfn(page) + pte_pfn(*pvmw.pte); pte 1431 mm/rmap.c if (huge_pmd_unshare(mm, &address, pvmw.pte)) { pte 1464 mm/rmap.c pteval = ptep_get_and_clear(mm, pvmw.address, pvmw.pte); pte 1475 mm/rmap.c set_pte_at(mm, pvmw.address, pvmw.pte, swp_pte); pte 1493 mm/rmap.c pvmw.pte)) { pte 1501 mm/rmap.c flush_cache_page(vma, address, pte_pfn(*pvmw.pte)); pte 1511 mm/rmap.c pteval = ptep_get_and_clear(mm, address, pvmw.pte); pte 1515 mm/rmap.c pteval = ptep_clear_flush(vma, address, pvmw.pte); pte 1530 mm/rmap.c pvmw.pte, pteval, pte 1534 mm/rmap.c set_pte_at(mm, address, pvmw.pte, pteval); pte 1558 mm/rmap.c set_pte_at(mm, address, pvmw.pte, pteval); pte 1574 mm/rmap.c set_pte_at(mm, address, pvmw.pte, swp_pte); pte 1610 mm/rmap.c set_pte_at(mm, address, pvmw.pte, pteval); pte 1618 mm/rmap.c set_pte_at(mm, address, pvmw.pte, pteval); pte 1624 mm/rmap.c set_pte_at(mm, address, pvmw.pte, pteval); pte 1640 mm/rmap.c set_pte_at(mm, address, pvmw.pte, swp_pte); pte 132 mm/sparse-vmemmap.c void __meminit vmemmap_verify(pte_t *pte, int node, pte 135 mm/sparse-vmemmap.c unsigned long pfn = pte_pfn(*pte); pte 145 mm/sparse-vmemmap.c pte_t *pte = pte_offset_kernel(pmd, addr); pte 146 mm/sparse-vmemmap.c if (pte_none(*pte)) { pte 152 mm/sparse-vmemmap.c set_pte_at(&init_mm, addr, pte, entry); pte 154 mm/sparse-vmemmap.c return pte; pte 224 mm/sparse-vmemmap.c pte_t *pte; pte 239 mm/sparse-vmemmap.c pte = vmemmap_pte_populate(pmd, addr, node); pte 240 mm/sparse-vmemmap.c if (!pte) pte 242 mm/sparse-vmemmap.c vmemmap_verify(pte, node, addr, addr + PAGE_SIZE); pte 648 mm/swap_state.c pte_t *pte, *orig_pte; pte 662 mm/swap_state.c orig_pte = pte = pte_offset_map(vmf->pmd, faddr); pte 663 mm/swap_state.c entry = pte_to_swp_entry(*pte); pte 697 mm/swap_state.c pte -= ra_info->offset; pte 699 mm/swap_state.c ra_info->ptes = pte; pte 703 mm/swap_state.c *tpte++ = *pte++; pte 728 mm/swap_state.c pte_t *pte, pentry; pte 739 mm/swap_state.c for (i = 0, pte = ra_info.ptes; i < ra_info.nr_pte; pte 740 mm/swap_state.c i++, pte++) { pte 741 mm/swap_state.c pentry = *pte; pte 1843 mm/swapfile.c static inline int pte_same_as_swp(pte_t pte, pte_t swp_pte) pte 1845 mm/swapfile.c return pte_same(pte_swp_clear_soft_dirty(pte), swp_pte); pte 1859 mm/swapfile.c pte_t *pte; pte 1873 mm/swapfile.c pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl); pte 1874 mm/swapfile.c if (unlikely(!pte_same_as_swp(*pte, swp_entry_to_pte(entry)))) { pte 1883 mm/swapfile.c set_pte_at(vma->vm_mm, addr, pte, pte 1900 mm/swapfile.c pte_unmap_unlock(pte, ptl); pte 1916 mm/swapfile.c pte_t *pte; pte 1923 mm/swapfile.c pte = pte_offset_map(pmd, addr); pte 1927 mm/swapfile.c if (!is_swap_pte(*pte)) pte 1930 mm/swapfile.c entry = pte_to_swp_entry(*pte); pte 1938 mm/swapfile.c pte_unmap(pte); pte 1968 mm/swapfile.c pte = pte_offset_map(pmd, addr); pte 1969 mm/swapfile.c } while (pte++, addr += PAGE_SIZE, addr != end); pte 1970 mm/swapfile.c pte_unmap(pte - 1); pte 66 mm/vmalloc.c pte_t *pte; pte 68 mm/vmalloc.c pte = pte_offset_kernel(pmd, addr); pte 70 mm/vmalloc.c pte_t ptent = ptep_get_and_clear(&init_mm, addr, pte); pte 72 mm/vmalloc.c } while (pte++, addr += PAGE_SIZE, addr != end); pte 141 mm/vmalloc.c pte_t *pte; pte 148 mm/vmalloc.c pte = pte_alloc_kernel(pmd, addr); pte 149 mm/vmalloc.c if (!pte) pte 154 mm/vmalloc.c if (WARN_ON(!pte_none(*pte))) pte 158 mm/vmalloc.c set_pte_at(&init_mm, addr, pte, mk_pte(page, prot)); pte 160 mm/vmalloc.c } while (pte++, addr += PAGE_SIZE, addr != end); pte 278 mm/vmalloc.c pte_t *ptep, pte; pte 310 mm/vmalloc.c pte = *ptep; pte 311 mm/vmalloc.c if (pte_present(pte)) pte 312 mm/vmalloc.c page = pte_page(pte); pte 3077 mm/vmalloc.c static int f(pte_t *pte, unsigned long addr, void *data) pte 3082 mm/vmalloc.c *(*p) = pte; pte 444 tools/lib/traceevent/plugins/plugin_kvm.c unsigned long pte = args[0]; pte 445 tools/lib/traceevent/plugins/plugin_kvm.c return pte & PT_WRITABLE_MASK; pte 191 tools/testing/selftests/kvm/lib/aarch64/processor.c uint64_t pte, *ptep; pte 196 tools/testing/selftests/kvm/lib/aarch64/processor.c for (pte = page; pte < page + ptrs_per_pte(vm) * 8; pte += 8) { pte 197 tools/testing/selftests/kvm/lib/aarch64/processor.c ptep = addr_gpa2hva(vm, pte); pte 200 tools/testing/selftests/kvm/lib/aarch64/processor.c printf("%*s%s: %lx: %lx at %p\n", indent, "", type[level], pte, *ptep, ptep); pte 158 tools/testing/selftests/kvm/lib/s390x/processor.c uint64_t *pte, ptea; pte 161 tools/testing/selftests/kvm/lib/s390x/processor.c pte = addr_gpa2hva(vm, ptea); pte 162 tools/testing/selftests/kvm/lib/s390x/processor.c if (*pte & PAGE_INVALID) pte 165 tools/testing/selftests/kvm/lib/s390x/processor.c indent, "", ptea, *pte); pte 322 tools/testing/selftests/kvm/lib/x86_64/processor.c struct pageTableEntry *pte; pte 323 tools/testing/selftests/kvm/lib/x86_64/processor.c pte = addr_gpa2hva(vm, pde[index[1]].address * vm->page_size); pte 324 tools/testing/selftests/kvm/lib/x86_64/processor.c pte[index[0]].address = paddr >> vm->page_shift; pte 325 tools/testing/selftests/kvm/lib/x86_64/processor.c pte[index[0]].writable = true; pte 326 tools/testing/selftests/kvm/lib/x86_64/processor.c pte[index[0]].present = 1; pte 348 tools/testing/selftests/kvm/lib/x86_64/processor.c struct pageTableEntry *pte, *pte_start; pte 401 tools/testing/selftests/kvm/lib/x86_64/processor.c pte = &pte_start[n4]; pte 402 tools/testing/selftests/kvm/lib/x86_64/processor.c if (!pte->present) pte 408 tools/testing/selftests/kvm/lib/x86_64/processor.c pte - pte_start, pte, pte 409 tools/testing/selftests/kvm/lib/x86_64/processor.c addr_hva2gpa(vm, pte), pte 410 tools/testing/selftests/kvm/lib/x86_64/processor.c (uint64_t) pte->address, pte 411 tools/testing/selftests/kvm/lib/x86_64/processor.c pte->writable, pte 412 tools/testing/selftests/kvm/lib/x86_64/processor.c pte->execute_disable, pte 413 tools/testing/selftests/kvm/lib/x86_64/processor.c pte->dirty, pte 548 tools/testing/selftests/kvm/lib/x86_64/processor.c struct pageTableEntry *pte; pte 572 tools/testing/selftests/kvm/lib/x86_64/processor.c pte = addr_gpa2hva(vm, pde[index[1]].address * vm->page_size); pte 573 tools/testing/selftests/kvm/lib/x86_64/processor.c if (!pte[index[0]].present) pte 576 tools/testing/selftests/kvm/lib/x86_64/processor.c return (pte[index[0]].address * vm->page_size) + (gva & 0xfffu); pte 459 tools/testing/selftests/kvm/lib/x86_64/vmx.c struct eptPageTableEntry *pte; pte 460 tools/testing/selftests/kvm/lib/x86_64/vmx.c pte = addr_gpa2hva(vm, pde[index[1]].address * vm->page_size); pte 461 tools/testing/selftests/kvm/lib/x86_64/vmx.c pte[index[0]].address = paddr >> vm->page_shift; pte 462 tools/testing/selftests/kvm/lib/x86_64/vmx.c pte[index[0]].writable = true; pte 463 tools/testing/selftests/kvm/lib/x86_64/vmx.c pte[index[0]].readable = true; pte 464 tools/testing/selftests/kvm/lib/x86_64/vmx.c pte[index[0]].executable = true; pte 470 tools/testing/selftests/kvm/lib/x86_64/vmx.c pte[index[0]].accessed = true; pte 471 tools/testing/selftests/kvm/lib/x86_64/vmx.c pte[index[0]].dirty = true; pte 72 virt/kvm/arm/mmu.c static void kvm_flush_dcache_pte(pte_t pte) pte 74 virt/kvm/arm/mmu.c __kvm_flush_dcache_pte(pte); pte 246 virt/kvm/arm/mmu.c pte_t *pte, *start_pte; pte 248 virt/kvm/arm/mmu.c start_pte = pte = pte_offset_kernel(pmd, addr); pte 250 virt/kvm/arm/mmu.c if (!pte_none(*pte)) { pte 251 virt/kvm/arm/mmu.c pte_t old_pte = *pte; pte 253 virt/kvm/arm/mmu.c kvm_set_pte(pte, __pte(0)); pte 260 virt/kvm/arm/mmu.c put_page(virt_to_page(pte)); pte 262 virt/kvm/arm/mmu.c } while (pte++, addr += PAGE_SIZE, addr != end); pte 368 virt/kvm/arm/mmu.c pte_t *pte; pte 370 virt/kvm/arm/mmu.c pte = pte_offset_kernel(pmd, addr); pte 372 virt/kvm/arm/mmu.c if (!pte_none(*pte) && !kvm_is_device_pfn(pte_pfn(*pte))) pte 373 virt/kvm/arm/mmu.c kvm_flush_dcache_pte(*pte); pte 374 virt/kvm/arm/mmu.c } while (pte++, addr += PAGE_SIZE, addr != end); pte 481 virt/kvm/arm/mmu.c pte_t *pte, *start_pte; pte 483 virt/kvm/arm/mmu.c start_pte = pte = pte_offset_kernel(pmd, addr); pte 485 virt/kvm/arm/mmu.c if (!pte_none(*pte)) { pte 486 virt/kvm/arm/mmu.c kvm_set_pte(pte, __pte(0)); pte 487 virt/kvm/arm/mmu.c put_page(virt_to_page(pte)); pte 489 virt/kvm/arm/mmu.c } while (pte++, addr += PAGE_SIZE, addr != end); pte 614 virt/kvm/arm/mmu.c pte_t *pte; pte 619 virt/kvm/arm/mmu.c pte = pte_offset_kernel(pmd, addr); pte 620 virt/kvm/arm/mmu.c kvm_set_pte(pte, kvm_pfn_pte(pfn, prot)); pte 621 virt/kvm/arm/mmu.c get_page(virt_to_page(pte)); pte 631 virt/kvm/arm/mmu.c pte_t *pte; pte 641 virt/kvm/arm/mmu.c pte = pte_alloc_one_kernel(NULL); pte 642 virt/kvm/arm/mmu.c if (!pte) { pte 646 virt/kvm/arm/mmu.c kvm_pmd_populate(pmd, pte); pte 1227 virt/kvm/arm/mmu.c pte_t *pte, old_pte; pte 1278 virt/kvm/arm/mmu.c pte = mmu_memory_cache_alloc(cache); pte 1279 virt/kvm/arm/mmu.c kvm_pmd_populate(pmd, pte); pte 1283 virt/kvm/arm/mmu.c pte = pte_offset_kernel(pmd, addr); pte 1285 virt/kvm/arm/mmu.c if (iomap && pte_present(*pte)) pte 1289 virt/kvm/arm/mmu.c old_pte = *pte; pte 1295 virt/kvm/arm/mmu.c kvm_set_pte(pte, __pte(0)); pte 1298 virt/kvm/arm/mmu.c get_page(virt_to_page(pte)); pte 1301 virt/kvm/arm/mmu.c kvm_set_pte(pte, *new_pte); pte 1306 virt/kvm/arm/mmu.c static int stage2_ptep_test_and_clear_young(pte_t *pte) pte 1308 virt/kvm/arm/mmu.c if (pte_young(*pte)) { pte 1309 virt/kvm/arm/mmu.c *pte = pte_mkold(*pte); pte 1315 virt/kvm/arm/mmu.c static int stage2_ptep_test_and_clear_young(pte_t *pte) pte 1317 virt/kvm/arm/mmu.c return __ptep_test_and_clear_young(pte); pte 1351 virt/kvm/arm/mmu.c pte_t pte = kvm_pfn_pte(pfn, PAGE_S2_DEVICE); pte 1354 virt/kvm/arm/mmu.c pte = kvm_s2pte_mkwrite(pte); pte 1362 virt/kvm/arm/mmu.c ret = stage2_set_pte(kvm, &cache, addr, &pte, pte 1431 virt/kvm/arm/mmu.c pte_t *pte; pte 1433 virt/kvm/arm/mmu.c pte = pte_offset_kernel(pmd, addr); pte 1435 virt/kvm/arm/mmu.c if (!pte_none(*pte)) { pte 1436 virt/kvm/arm/mmu.c if (!kvm_s2pte_readonly(pte)) pte 1437 virt/kvm/arm/mmu.c kvm_set_s2pte_readonly(pte); pte 1439 virt/kvm/arm/mmu.c } while (pte++, addr += PAGE_SIZE, addr != end); pte 1865 virt/kvm/arm/mmu.c pte_t *pte; pte 1873 virt/kvm/arm/mmu.c if (!stage2_get_leaf_entry(vcpu->kvm, fault_ipa, &pud, &pmd, &pte)) pte 1885 virt/kvm/arm/mmu.c *pte = pte_mkyoung(*pte); /* Just a page... */ pte 1886 virt/kvm/arm/mmu.c pfn = pte_pfn(*pte); pte 2064 virt/kvm/arm/mmu.c pte_t *pte = (pte_t *)data; pte 2074 virt/kvm/arm/mmu.c stage2_set_pte(kvm, NULL, gpa, pte, 0); pte 2079 virt/kvm/arm/mmu.c int kvm_set_spte_hva(struct kvm *kvm, unsigned long hva, pte_t pte) pte 2082 virt/kvm/arm/mmu.c kvm_pfn_t pfn = pte_pfn(pte); pte 2105 virt/kvm/arm/mmu.c pte_t *pte; pte 2108 virt/kvm/arm/mmu.c if (!stage2_get_leaf_entry(kvm, gpa, &pud, &pmd, &pte)) pte 2116 virt/kvm/arm/mmu.c return stage2_ptep_test_and_clear_young(pte); pte 2123 virt/kvm/arm/mmu.c pte_t *pte; pte 2126 virt/kvm/arm/mmu.c if (!stage2_get_leaf_entry(kvm, gpa, &pud, &pmd, &pte)) pte 2134 virt/kvm/arm/mmu.c return pte_young(*pte); pte 398 virt/kvm/kvm_main.c pte_t pte) pte 407 virt/kvm/kvm_main.c if (kvm_set_spte_hva(kvm, address, pte))