pte                77 arch/alpha/include/asm/mmzone.h #define pte_pfn(pte)		(pte_val(pte) >> 32)
pte                81 arch/alpha/include/asm/mmzone.h 	pte_t pte;                                                           \
pte                85 arch/alpha/include/asm/mmzone.h 	pte_val(pte) = pfn | pgprot_val(pgprot);			     \
pte                87 arch/alpha/include/asm/mmzone.h 	pte;								     \
pte                31 arch/alpha/include/asm/page.h typedef struct { unsigned long pte; } pte_t;
pte                36 arch/alpha/include/asm/page.h #define pte_val(x)	((x).pte)
pte                17 arch/alpha/include/asm/pgalloc.h pmd_populate(struct mm_struct *mm, pmd_t *pmd, pgtable_t pte)
pte                19 arch/alpha/include/asm/pgalloc.h 	pmd_set(pmd, (pte_t *)(page_to_pa(pte) + PAGE_OFFSET));
pte                24 arch/alpha/include/asm/pgalloc.h pmd_populate_kernel(struct mm_struct *mm, pmd_t *pmd, pte_t *pte)
pte                26 arch/alpha/include/asm/pgalloc.h 	pmd_set(pmd, pte);
pte               209 arch/alpha/include/asm/pgtable.h #define pte_pfn(pte)	(pte_val(pte) >> 32)
pte               210 arch/alpha/include/asm/pgtable.h #define pte_page(pte)	pfn_to_page(pte_pfn(pte))
pte               213 arch/alpha/include/asm/pgtable.h 	pte_t pte;							\
pte               215 arch/alpha/include/asm/pgtable.h 	pte_val(pte) = (page_to_pfn(page) << 32) | pgprot_val(pgprot);	\
pte               216 arch/alpha/include/asm/pgtable.h 	pte;								\
pte               221 arch/alpha/include/asm/pgtable.h { pte_t pte; pte_val(pte) = (PHYS_TWIDDLE(physpfn) << 32) | pgprot_val(pgprot); return pte; }
pte               223 arch/alpha/include/asm/pgtable.h extern inline pte_t pte_modify(pte_t pte, pgprot_t newprot)
pte               224 arch/alpha/include/asm/pgtable.h { pte_val(pte) = (pte_val(pte) & _PAGE_CHG_MASK) | pgprot_val(newprot); return pte; }
pte               247 arch/alpha/include/asm/pgtable.h extern inline int pte_none(pte_t pte)		{ return !pte_val(pte); }
pte               248 arch/alpha/include/asm/pgtable.h extern inline int pte_present(pte_t pte)	{ return pte_val(pte) & _PAGE_VALID; }
pte               268 arch/alpha/include/asm/pgtable.h extern inline int pte_write(pte_t pte)		{ return !(pte_val(pte) & _PAGE_FOW); }
pte               269 arch/alpha/include/asm/pgtable.h extern inline int pte_dirty(pte_t pte)		{ return pte_val(pte) & _PAGE_DIRTY; }
pte               270 arch/alpha/include/asm/pgtable.h extern inline int pte_young(pte_t pte)		{ return pte_val(pte) & _PAGE_ACCESSED; }
pte               271 arch/alpha/include/asm/pgtable.h extern inline int pte_special(pte_t pte)	{ return 0; }
pte               273 arch/alpha/include/asm/pgtable.h extern inline pte_t pte_wrprotect(pte_t pte)	{ pte_val(pte) |= _PAGE_FOW; return pte; }
pte               274 arch/alpha/include/asm/pgtable.h extern inline pte_t pte_mkclean(pte_t pte)	{ pte_val(pte) &= ~(__DIRTY_BITS); return pte; }
pte               275 arch/alpha/include/asm/pgtable.h extern inline pte_t pte_mkold(pte_t pte)	{ pte_val(pte) &= ~(__ACCESS_BITS); return pte; }
pte               276 arch/alpha/include/asm/pgtable.h extern inline pte_t pte_mkwrite(pte_t pte)	{ pte_val(pte) &= ~_PAGE_FOW; return pte; }
pte               277 arch/alpha/include/asm/pgtable.h extern inline pte_t pte_mkdirty(pte_t pte)	{ pte_val(pte) |= __DIRTY_BITS; return pte; }
pte               278 arch/alpha/include/asm/pgtable.h extern inline pte_t pte_mkyoung(pte_t pte)	{ pte_val(pte) |= __ACCESS_BITS; return pte; }
pte               279 arch/alpha/include/asm/pgtable.h extern inline pte_t pte_mkspecial(pte_t pte)	{ return pte; }
pte               321 arch/alpha/include/asm/pgtable.h #define pte_unmap(pte)			do { } while (0)
pte               339 arch/alpha/include/asm/pgtable.h { pte_t pte; pte_val(pte) = (type << 32) | (offset << 40); return pte; }
pte               344 arch/alpha/include/asm/pgtable.h #define __pte_to_swp_entry(pte)	((swp_entry_t) { pte_val(pte) })
pte                 7 arch/alpha/include/asm/tlb.h #define __pte_free_tlb(tlb, pte, address)		pte_free((tlb)->mm, pte)
pte               330 arch/alpha/kernel/core_cia.c 	unsigned long *ppte, pte;
pte               338 arch/alpha/kernel/core_cia.c 	pte = (virt_to_phys(ppte) >> (PAGE_SHIFT - 1)) | 1;
pte               341 arch/alpha/kernel/core_cia.c 		ppte[i] = pte;
pte               319 arch/alpha/kernel/core_irongate.c 	u32 *mmio_regs, *gatt_pages, *cur_gatt, pte;
pte               371 arch/alpha/kernel/core_irongate.c 		pte = cur_gatt[GET_GATT_OFF(baddr)] & ~1;
pte               373 arch/alpha/kernel/core_irongate.c 		       cur_gatt, pte);
pte               388 arch/alpha/kernel/core_irongate.c 		pte = cur_gatt[GET_GATT_OFF(baddr)] & ~1;
pte               391 arch/alpha/kernel/core_irongate.c 					     pte, PAGE_SIZE, 0)) {
pte              1043 arch/alpha/kernel/core_marvel.c 	unsigned long pte;
pte              1051 arch/alpha/kernel/core_marvel.c 	pte = aper->arena->ptes[baddr >> PAGE_SHIFT];
pte              1052 arch/alpha/kernel/core_marvel.c 	if (!(pte & 1)) {
pte              1056 arch/alpha/kernel/core_marvel.c 	return (pte >> 1) << PAGE_SHIFT;
pte               704 arch/alpha/kernel/core_titan.c 	unsigned long pte;
pte               712 arch/alpha/kernel/core_titan.c 	pte = aper->arena->ptes[baddr >> PAGE_SHIFT];
pte               713 arch/alpha/kernel/core_titan.c 	if (!(pte & 1)) {
pte               718 arch/alpha/kernel/core_titan.c 	return (pte >> 1) << PAGE_SHIFT;
pte               237 arch/arc/include/asm/arcregs.h 	unsigned int pad:6, fbe:2, tqe:2, ts:4, ft:1, rse:2, pte:3, bce:3, ver:8;
pte               239 arch/arc/include/asm/arcregs.h 	unsigned int ver:8, bce:3, pte:3, rse:2, ft:1, ts:4, tqe:2, fbe:2, pad:6;
pte                19 arch/arc/include/asm/hugepage.h static inline pmd_t pte_pmd(pte_t pte)
pte                21 arch/arc/include/asm/hugepage.h 	return __pmd(pte_val(pte));
pte                32 arch/arc/include/asm/page.h 	unsigned long long pte;
pte                34 arch/arc/include/asm/page.h 	unsigned long pte;
pte                44 arch/arc/include/asm/page.h #define pte_val(x)      ((x).pte)
pte                36 arch/arc/include/asm/pgalloc.h pmd_populate_kernel(struct mm_struct *mm, pmd_t *pmd, pte_t *pte)
pte                38 arch/arc/include/asm/pgalloc.h 	pmd_set(pmd, pte);
pte                92 arch/arc/include/asm/pgalloc.h 	pte_t *pte;
pte                94 arch/arc/include/asm/pgalloc.h 	pte = (pte_t *) __get_free_pages(GFP_KERNEL | __GFP_ZERO,
pte                97 arch/arc/include/asm/pgalloc.h 	return pte;
pte               119 arch/arc/include/asm/pgalloc.h static inline void pte_free_kernel(struct mm_struct *mm, pte_t *pte)
pte               121 arch/arc/include/asm/pgalloc.h 	free_pages((unsigned long)pte, __get_order_pte()); /* takes phy addr */
pte               130 arch/arc/include/asm/pgalloc.h #define __pte_free_tlb(tlb, pte, addr)  pte_free((tlb)->mm, pte)
pte               252 arch/arc/include/asm/pgtable.h #define pte_unmap(pte)		do { } while (0)
pte               253 arch/arc/include/asm/pgtable.h #define pte_unmap_nested(pte)		do { } while (0)
pte               279 arch/arc/include/asm/pgtable.h #define pte_page(pte)		pfn_to_page(pte_pfn(pte))
pte               284 arch/arc/include/asm/pgtable.h #define pte_pfn(pte)		(pte_val(pte) >> PAGE_SHIFT)
pte               299 arch/arc/include/asm/pgtable.h #define pte_read(pte)		(pte_val(pte) & _PAGE_READ)
pte               300 arch/arc/include/asm/pgtable.h #define pte_write(pte)		(pte_val(pte) & _PAGE_WRITE)
pte               301 arch/arc/include/asm/pgtable.h #define pte_dirty(pte)		(pte_val(pte) & _PAGE_DIRTY)
pte               302 arch/arc/include/asm/pgtable.h #define pte_young(pte)		(pte_val(pte) & _PAGE_ACCESSED)
pte               303 arch/arc/include/asm/pgtable.h #define pte_special(pte)	(pte_val(pte) & _PAGE_SPECIAL)
pte               306 arch/arc/include/asm/pgtable.h 	static inline pte_t pte_##fn(pte_t pte) { pte_val(pte) op; return pte; }
pte               320 arch/arc/include/asm/pgtable.h static inline pte_t pte_modify(pte_t pte, pgprot_t newprot)
pte               322 arch/arc/include/asm/pgtable.h 	return __pte((pte_val(pte) & _PAGE_CHG_MASK) | pgprot_val(newprot));
pte               380 arch/arc/include/asm/pgtable.h #define __pte_to_swp_entry(pte)	((swp_entry_t) { pte_val(pte) })
pte               223 arch/arc/kernel/setup.c 		cpu->bpu.num_pred = 2048 << bpu.pte;
pte               659 arch/arc/mm/tlb.c 	pte_t pte = __pte(pmd_val(*pmd));
pte               660 arch/arc/mm/tlb.c 	update_mmu_cache(vma, addr, &pte);
pte               270 arch/arm/include/asm/kvm_host.h int kvm_set_spte_hva(struct kvm *kvm, unsigned long hva, pte_t pte);
pte               134 arch/arm/include/asm/kvm_mmu.h static inline pte_t kvm_s2pte_mkwrite(pte_t pte)
pte               136 arch/arm/include/asm/kvm_mmu.h 	pte_val(pte) |= L_PTE_S2_RDWR;
pte               137 arch/arm/include/asm/kvm_mmu.h 	return pte;
pte               146 arch/arm/include/asm/kvm_mmu.h static inline pte_t kvm_s2pte_mkexec(pte_t pte)
pte               148 arch/arm/include/asm/kvm_mmu.h 	pte_val(pte) &= ~L_PTE_XN;
pte               149 arch/arm/include/asm/kvm_mmu.h 	return pte;
pte               158 arch/arm/include/asm/kvm_mmu.h static inline void kvm_set_s2pte_readonly(pte_t *pte)
pte               160 arch/arm/include/asm/kvm_mmu.h 	pte_val(*pte) = (pte_val(*pte) & ~L_PTE_S2_RDWR) | L_PTE_S2_RDONLY;
pte               163 arch/arm/include/asm/kvm_mmu.h static inline bool kvm_s2pte_readonly(pte_t *pte)
pte               165 arch/arm/include/asm/kvm_mmu.h 	return (pte_val(*pte) & L_PTE_S2_RDWR) == L_PTE_S2_RDONLY;
pte               168 arch/arm/include/asm/kvm_mmu.h static inline bool kvm_s2pte_exec(pte_t *pte)
pte               170 arch/arm/include/asm/kvm_mmu.h 	return !(pte_val(*pte) & L_PTE_XN);
pte               301 arch/arm/include/asm/kvm_mmu.h static inline void __kvm_flush_dcache_pte(pte_t pte)
pte               303 arch/arm/include/asm/kvm_mmu.h 	void *va = kmap_atomic(pte_page(pte));
pte                48 arch/arm/include/asm/pgalloc.h #define pud_populate(mm,pmd,pte)	BUG()
pte                55 arch/arm/include/asm/pgalloc.h static inline void clean_pte_table(pte_t *pte)
pte                57 arch/arm/include/asm/pgalloc.h 	clean_dcache_area(pte + PTE_HWTABLE_PTRS, PTE_HWTABLE_SIZE);
pte                84 arch/arm/include/asm/pgalloc.h 	pte_t *pte = __pte_alloc_one_kernel(mm);
pte                86 arch/arm/include/asm/pgalloc.h 	if (pte)
pte                87 arch/arm/include/asm/pgalloc.h 		clean_pte_table(pte);
pte                89 arch/arm/include/asm/pgalloc.h 	return pte;
pte               101 arch/arm/include/asm/pgalloc.h 	struct page *pte;
pte               103 arch/arm/include/asm/pgalloc.h 	pte = __pte_alloc_one(mm, GFP_PGTABLE_USER | PGTABLE_HIGHMEM);
pte               104 arch/arm/include/asm/pgalloc.h 	if (!pte)
pte               106 arch/arm/include/asm/pgalloc.h 	if (!PageHighMem(pte))
pte               107 arch/arm/include/asm/pgalloc.h 		clean_pte_table(page_address(pte));
pte               108 arch/arm/include/asm/pgalloc.h 	return pte;
pte               111 arch/arm/include/asm/pgalloc.h static inline void __pmd_populate(pmd_t *pmdp, phys_addr_t pte,
pte               114 arch/arm/include/asm/pgalloc.h 	pmdval_t pmdval = (pte + PTE_HWTABLE_OFF) | prot;
pte                21 arch/arm/include/asm/pgtable-2level-types.h typedef struct { pteval_t pte; } pte_t;
pte                26 arch/arm/include/asm/pgtable-2level-types.h #define pte_val(x)      ((x).pte)
pte               212 arch/arm/include/asm/pgtable-2level.h #define set_pte_ext(ptep,pte,ext) cpu_set_pte_ext(ptep,pte,ext)
pte               213 arch/arm/include/asm/pgtable-2level.h #define pte_special(pte)	(0)
pte               214 arch/arm/include/asm/pgtable-2level.h static inline pte_t pte_mkspecial(pte_t pte) { return pte; }
pte                24 arch/arm/include/asm/pgtable-3level-types.h typedef struct { pteval_t pte; } pte_t;
pte                29 arch/arm/include/asm/pgtable-3level-types.h #define pte_val(x)      ((x).pte)
pte               193 arch/arm/include/asm/pgtable-3level.h #define set_pte_ext(ptep,pte,ext) cpu_set_pte_ext(ptep,__pte(pte_val(pte)|(ext)))
pte               195 arch/arm/include/asm/pgtable-3level.h #define pte_huge(pte)		(pte_val(pte) && !(pte_val(pte) & PTE_TABLE_BIT))
pte               196 arch/arm/include/asm/pgtable-3level.h #define pte_mkhuge(pte)		(__pte(pte_val(pte) & ~PTE_TABLE_BIT))
pte               204 arch/arm/include/asm/pgtable-3level.h #define pte_special(pte)	(pte_isset((pte), L_PTE_SPECIAL))
pte               205 arch/arm/include/asm/pgtable-3level.h static inline pte_t pte_mkspecial(pte_t pte)
pte               207 arch/arm/include/asm/pgtable-3level.h 	pte_val(pte) |= L_PTE_SPECIAL;
pte               208 arch/arm/include/asm/pgtable-3level.h 	return pte;
pte                48 arch/arm/include/asm/pgtable-nommu.h #define __pte_to_swp_entry(pte)	((swp_entry_t) { pte_val(pte) })
pte                53 arch/arm/include/asm/pgtable.h #define pte_ERROR(pte)		__pte_error(__FILE__, __LINE__, pte)
pte               198 arch/arm/include/asm/pgtable.h #define __pte_unmap(pte)	do { } while (0)
pte               201 arch/arm/include/asm/pgtable.h #define __pte_unmap(pte)	kunmap_atomic(pte)
pte               209 arch/arm/include/asm/pgtable.h #define pte_unmap(pte)			__pte_unmap(pte)
pte               211 arch/arm/include/asm/pgtable.h #define pte_pfn(pte)		((pte_val(pte) & PHYS_MASK) >> PAGE_SHIFT)
pte               214 arch/arm/include/asm/pgtable.h #define pte_page(pte)		pfn_to_page(pte_pfn(pte))
pte               219 arch/arm/include/asm/pgtable.h #define pte_isset(pte, val)	((u32)(val) == (val) ? pte_val(pte) & (val) \
pte               220 arch/arm/include/asm/pgtable.h 						: !!(pte_val(pte) & (val)))
pte               221 arch/arm/include/asm/pgtable.h #define pte_isclear(pte, val)	(!(pte_val(pte) & (val)))
pte               223 arch/arm/include/asm/pgtable.h #define pte_none(pte)		(!pte_val(pte))
pte               224 arch/arm/include/asm/pgtable.h #define pte_present(pte)	(pte_isset((pte), L_PTE_PRESENT))
pte               225 arch/arm/include/asm/pgtable.h #define pte_valid(pte)		(pte_isset((pte), L_PTE_VALID))
pte               226 arch/arm/include/asm/pgtable.h #define pte_accessible(mm, pte)	(mm_tlb_flush_pending(mm) ? pte_present(pte) : pte_valid(pte))
pte               227 arch/arm/include/asm/pgtable.h #define pte_write(pte)		(pte_isclear((pte), L_PTE_RDONLY))
pte               228 arch/arm/include/asm/pgtable.h #define pte_dirty(pte)		(pte_isset((pte), L_PTE_DIRTY))
pte               229 arch/arm/include/asm/pgtable.h #define pte_young(pte)		(pte_isset((pte), L_PTE_YOUNG))
pte               230 arch/arm/include/asm/pgtable.h #define pte_exec(pte)		(pte_isclear((pte), L_PTE_XN))
pte               232 arch/arm/include/asm/pgtable.h #define pte_valid_user(pte)	\
pte               233 arch/arm/include/asm/pgtable.h 	(pte_valid(pte) && pte_isset((pte), L_PTE_USER) && pte_young(pte))
pte               235 arch/arm/include/asm/pgtable.h static inline bool pte_access_permitted(pte_t pte, bool write)
pte               243 arch/arm/include/asm/pgtable.h 	return (pte_val(pte) & mask) == needed;
pte               269 arch/arm/include/asm/pgtable.h static inline pte_t clear_pte_bit(pte_t pte, pgprot_t prot)
pte               271 arch/arm/include/asm/pgtable.h 	pte_val(pte) &= ~pgprot_val(prot);
pte               272 arch/arm/include/asm/pgtable.h 	return pte;
pte               275 arch/arm/include/asm/pgtable.h static inline pte_t set_pte_bit(pte_t pte, pgprot_t prot)
pte               277 arch/arm/include/asm/pgtable.h 	pte_val(pte) |= pgprot_val(prot);
pte               278 arch/arm/include/asm/pgtable.h 	return pte;
pte               281 arch/arm/include/asm/pgtable.h static inline pte_t pte_wrprotect(pte_t pte)
pte               283 arch/arm/include/asm/pgtable.h 	return set_pte_bit(pte, __pgprot(L_PTE_RDONLY));
pte               286 arch/arm/include/asm/pgtable.h static inline pte_t pte_mkwrite(pte_t pte)
pte               288 arch/arm/include/asm/pgtable.h 	return clear_pte_bit(pte, __pgprot(L_PTE_RDONLY));
pte               291 arch/arm/include/asm/pgtable.h static inline pte_t pte_mkclean(pte_t pte)
pte               293 arch/arm/include/asm/pgtable.h 	return clear_pte_bit(pte, __pgprot(L_PTE_DIRTY));
pte               296 arch/arm/include/asm/pgtable.h static inline pte_t pte_mkdirty(pte_t pte)
pte               298 arch/arm/include/asm/pgtable.h 	return set_pte_bit(pte, __pgprot(L_PTE_DIRTY));
pte               301 arch/arm/include/asm/pgtable.h static inline pte_t pte_mkold(pte_t pte)
pte               303 arch/arm/include/asm/pgtable.h 	return clear_pte_bit(pte, __pgprot(L_PTE_YOUNG));
pte               306 arch/arm/include/asm/pgtable.h static inline pte_t pte_mkyoung(pte_t pte)
pte               308 arch/arm/include/asm/pgtable.h 	return set_pte_bit(pte, __pgprot(L_PTE_YOUNG));
pte               311 arch/arm/include/asm/pgtable.h static inline pte_t pte_mkexec(pte_t pte)
pte               313 arch/arm/include/asm/pgtable.h 	return clear_pte_bit(pte, __pgprot(L_PTE_XN));
pte               316 arch/arm/include/asm/pgtable.h static inline pte_t pte_mknexec(pte_t pte)
pte               318 arch/arm/include/asm/pgtable.h 	return set_pte_bit(pte, __pgprot(L_PTE_XN));
pte               321 arch/arm/include/asm/pgtable.h static inline pte_t pte_modify(pte_t pte, pgprot_t newprot)
pte               325 arch/arm/include/asm/pgtable.h 	pte_val(pte) = (pte_val(pte) & ~mask) | (pgprot_val(newprot) & mask);
pte               326 arch/arm/include/asm/pgtable.h 	return pte;
pte               349 arch/arm/include/asm/pgtable.h #define __pte_to_swp_entry(pte)	((swp_entry_t) { pte_val(pte) })
pte                70 arch/arm/include/asm/proc-fns.h 	void (*set_pte_ext)(pte_t *ptep, pte_t pte);
pte                72 arch/arm/include/asm/proc-fns.h 	void (*set_pte_ext)(pte_t *ptep, pte_t pte, unsigned int ext);
pte                92 arch/arm/include/asm/proc-fns.h extern void cpu_set_pte_ext(pte_t *ptep, pte_t pte);
pte                94 arch/arm/include/asm/proc-fns.h extern void cpu_set_pte_ext(pte_t *ptep, pte_t pte, unsigned int ext);
pte                45 arch/arm/include/asm/tlb.h __pte_free_tlb(struct mmu_gather *tlb, pgtable_t pte, unsigned long addr)
pte                47 arch/arm/include/asm/tlb.h 	pgtable_pte_page_dtor(pte);
pte                58 arch/arm/include/asm/tlb.h 	tlb_remove_table(tlb, pte);
pte                14 arch/arm/kernel/efi.c 	pte_t pte = *ptep;
pte                17 arch/arm/kernel/efi.c 		pte = set_pte_bit(pte, __pgprot(L_PTE_RDONLY));
pte                19 arch/arm/kernel/efi.c 		pte = set_pte_bit(pte, __pgprot(L_PTE_XN));
pte                20 arch/arm/kernel/efi.c 	set_pte_ext(ptep, pte, PTE_EXT_NG);
pte               744 arch/arm/kernel/traps.c void __pte_error(const char *file, int line, pte_t pte)
pte               746 arch/arm/kernel/traps.c 	pr_err("%s:%d: bad pte %08llx.\n", file, line, (long long)pte_val(pte));
pte                28 arch/arm/lib/uaccess_with_memcpy.c 	pte_t *pte;
pte                71 arch/arm/lib/uaccess_with_memcpy.c 	pte = pte_offset_map_lock(current->mm, pmd, addr, &ptl);
pte                72 arch/arm/lib/uaccess_with_memcpy.c 	if (unlikely(!pte_present(*pte) || !pte_young(*pte) ||
pte                73 arch/arm/lib/uaccess_with_memcpy.c 	    !pte_write(*pte) || !pte_dirty(*pte))) {
pte                74 arch/arm/lib/uaccess_with_memcpy.c 		pte_unmap_unlock(pte, ptl);
pte                78 arch/arm/lib/uaccess_with_memcpy.c 	*ptep = pte;
pte               101 arch/arm/lib/uaccess_with_memcpy.c 		pte_t *pte;
pte               105 arch/arm/lib/uaccess_with_memcpy.c 		while (!pin_page_for_write(to, &pte, &ptl)) {
pte               125 arch/arm/lib/uaccess_with_memcpy.c 		if (pte)
pte               126 arch/arm/lib/uaccess_with_memcpy.c 			pte_unmap_unlock(pte, ptl);
pte               170 arch/arm/lib/uaccess_with_memcpy.c 		pte_t *pte;
pte               174 arch/arm/lib/uaccess_with_memcpy.c 		while (!pin_page_for_write(addr, &pte, &ptl)) {
pte               191 arch/arm/lib/uaccess_with_memcpy.c 		if (pte)
pte               192 arch/arm/lib/uaccess_with_memcpy.c 			pte_unmap_unlock(pte, ptl);
pte               464 arch/arm/mm/dma-mapping.c static int __dma_update_pte(pte_t *pte, unsigned long addr, void *data)
pte               469 arch/arm/mm/dma-mapping.c 	set_pte_ext(pte, mk_pte(page, prot), 0);
pte               305 arch/arm/mm/dump.c 	pte_t *pte = pte_offset_kernel(pmd, 0);
pte               309 arch/arm/mm/dump.c 	for (i = 0; i < PTRS_PER_PTE; i++, pte++) {
pte               311 arch/arm/mm/dump.c 		note_page(st, addr, 4, pte_val(*pte), domain);
pte                96 arch/arm/mm/fault-armv.c 	pte_t *pte;
pte               117 arch/arm/mm/fault-armv.c 	pte = pte_offset_map(pmd, address);
pte               120 arch/arm/mm/fault-armv.c 	ret = do_adjust_pte(vma, address, pfn, pte);
pte               123 arch/arm/mm/fault-armv.c 	pte_unmap(pte);
pte                48 arch/arm/mm/fault.c 		pte_t *pte;
pte                86 arch/arm/mm/fault.c 		pte = pte_offset_map(pmd, addr);
pte                87 arch/arm/mm/fault.c 		pr_cont(", *pte=%08llx", (long long)pte_val(*pte));
pte                90 arch/arm/mm/fault.c 		       (long long)pte_val(pte[PTE_HWTABLE_PTRS]));
pte                92 arch/arm/mm/fault.c 		pte_unmap(pte);
pte                18 arch/arm/mm/highmem.c static inline void set_fixmap_pte(int idx, pte_t pte)
pte                23 arch/arm/mm/highmem.c 	set_pte_ext(ptep, pte, 0);
pte                26 arch/arm/mm/mm.h static inline void set_top_pte(unsigned long va, pte_t pte)
pte                29 arch/arm/mm/mm.h 	set_pte_ext(ptep, pte, 0);
pte                77 arch/arm/mm/mmu.c 	pteval_t	pte;
pte                94 arch/arm/mm/mmu.c 		.pte		= L_PTE_MT_UNCACHED,
pte               100 arch/arm/mm/mmu.c 		.pte		= L_PTE_MT_BUFFERABLE,
pte               106 arch/arm/mm/mmu.c 		.pte		= L_PTE_MT_WRITETHROUGH,
pte               112 arch/arm/mm/mmu.c 		.pte		= L_PTE_MT_WRITEBACK,
pte               118 arch/arm/mm/mmu.c 		.pte		= L_PTE_MT_WRITEALLOC,
pte               409 arch/arm/mm/mmu.c 	pte_t *pte = pte_offset_fixmap(pmd_off_k(vaddr), vaddr);
pte               422 arch/arm/mm/mmu.c 		set_pte_at(NULL, vaddr, pte,
pte               425 arch/arm/mm/mmu.c 		pte_clear(NULL, vaddr, pte);
pte               560 arch/arm/mm/mmu.c 	vecs_pgprot = kern_pgprot = user_pgprot = cp->pte;
pte               744 arch/arm/mm/mmu.c 		pte_t *pte = alloc(PTE_HWTABLE_OFF + PTE_HWTABLE_SIZE);
pte               745 arch/arm/mm/mmu.c 		__pmd_populate(pmd, __pa(pte), prot);
pte               763 arch/arm/mm/mmu.c 	pte_t *pte = arm_pte_alloc(pmd, addr, type->prot_l1, alloc);
pte               765 arch/arm/mm/mmu.c 		set_pte_ext(pte, pfn_pte(pfn, __pgprot(type->prot_pte)),
pte               768 arch/arm/mm/mmu.c 	} while (pte++, addr += PAGE_SIZE, addr != end);
pte              1620 arch/arm/mm/mmu.c 		pte_t *pte;
pte              1624 arch/arm/mm/mmu.c 		pte = pte_offset_early_fixmap(pmd_off_k(map.virtual), map.virtual);
pte              1627 arch/arm/mm/mmu.c 		if (pte_none(*pte) ||
pte              1628 arch/arm/mm/mmu.c 		    (pte_val(*pte) & L_PTE_MT_MASK) != L_PTE_MT_DEV_SHARED)
pte              1631 arch/arm/mm/mmu.c 		map.pfn = pte_pfn(*pte);
pte                20 arch/arm/mm/pageattr.c 	pte_t pte = *ptep;
pte                22 arch/arm/mm/pageattr.c 	pte = clear_pte_bit(pte, cdata->clear_mask);
pte                23 arch/arm/mm/pageattr.c 	pte = set_pte_bit(pte, cdata->set_mask);
pte                25 arch/arm/mm/pageattr.c 	set_pte_ext(ptep, pte, 0);
pte               121 arch/arm/mm/pgd.c 	pgtable_t pte;
pte               138 arch/arm/mm/pgd.c 	pte = pmd_pgtable(*pmd);
pte               140 arch/arm/mm/pgd.c 	pte_free(mm, pte);
pte               574 arch/arm64/include/asm/assembler.h 	.macro	phys_to_pte, pte, phys
pte               587 arch/arm64/include/asm/assembler.h 	.macro	pte_to_phys, phys, pte
pte                42 arch/arm64/include/asm/hugetlb.h 			    pte_t *ptep, pte_t pte);
pte                46 arch/arm64/include/asm/hugetlb.h 				      pte_t pte, int dirty);
pte                60 arch/arm64/include/asm/hugetlb.h 				 pte_t *ptep, pte_t pte, unsigned long sz);
pte               431 arch/arm64/include/asm/kvm_host.h int kvm_set_spte_hva(struct kvm *kvm, unsigned long hva, pte_t pte);
pte               188 arch/arm64/include/asm/kvm_mmu.h static inline pte_t kvm_s2pte_mkwrite(pte_t pte)
pte               190 arch/arm64/include/asm/kvm_mmu.h 	pte_val(pte) |= PTE_S2_RDWR;
pte               191 arch/arm64/include/asm/kvm_mmu.h 	return pte;
pte               206 arch/arm64/include/asm/kvm_mmu.h static inline pte_t kvm_s2pte_mkexec(pte_t pte)
pte               208 arch/arm64/include/asm/kvm_mmu.h 	pte_val(pte) &= ~PTE_S2_XN;
pte               209 arch/arm64/include/asm/kvm_mmu.h 	return pte;
pte               341 arch/arm64/include/asm/kvm_mmu.h static inline void __kvm_flush_dcache_pte(pte_t pte)
pte               344 arch/arm64/include/asm/kvm_mmu.h 		struct page *page = pte_page(pte);
pte                22 arch/arm64/include/asm/pgtable-types.h typedef struct { pteval_t pte; } pte_t;
pte                23 arch/arm64/include/asm/pgtable-types.h #define pte_val(x)	((x).pte)
pte                50 arch/arm64/include/asm/pgtable.h #define pte_ERROR(pte)		__pte_error(__FILE__, __LINE__, pte_val(pte))
pte                57 arch/arm64/include/asm/pgtable.h #define __pte_to_phys(pte)	\
pte                58 arch/arm64/include/asm/pgtable.h 	((pte_val(pte) & PTE_ADDR_LOW) | ((pte_val(pte) & PTE_ADDR_HIGH) << 36))
pte                61 arch/arm64/include/asm/pgtable.h #define __pte_to_phys(pte)	(pte_val(pte) & PTE_ADDR_MASK)
pte                65 arch/arm64/include/asm/pgtable.h #define pte_pfn(pte)		(__pte_to_phys(pte) >> PAGE_SHIFT)
pte                69 arch/arm64/include/asm/pgtable.h #define pte_none(pte)		(!pte_val(pte))
pte                71 arch/arm64/include/asm/pgtable.h #define pte_page(pte)		(pfn_to_page(pte_pfn(pte)))
pte                76 arch/arm64/include/asm/pgtable.h #define pte_present(pte)	(!!(pte_val(pte) & (PTE_VALID | PTE_PROT_NONE)))
pte                77 arch/arm64/include/asm/pgtable.h #define pte_young(pte)		(!!(pte_val(pte) & PTE_AF))
pte                78 arch/arm64/include/asm/pgtable.h #define pte_special(pte)	(!!(pte_val(pte) & PTE_SPECIAL))
pte                79 arch/arm64/include/asm/pgtable.h #define pte_write(pte)		(!!(pte_val(pte) & PTE_WRITE))
pte                80 arch/arm64/include/asm/pgtable.h #define pte_user_exec(pte)	(!(pte_val(pte) & PTE_UXN))
pte                81 arch/arm64/include/asm/pgtable.h #define pte_cont(pte)		(!!(pte_val(pte) & PTE_CONT))
pte                82 arch/arm64/include/asm/pgtable.h #define pte_devmap(pte)		(!!(pte_val(pte) & PTE_DEVMAP))
pte                94 arch/arm64/include/asm/pgtable.h #define pte_hw_dirty(pte)	(pte_write(pte) && !(pte_val(pte) & PTE_RDONLY))
pte                95 arch/arm64/include/asm/pgtable.h #define pte_sw_dirty(pte)	(!!(pte_val(pte) & PTE_DIRTY))
pte                96 arch/arm64/include/asm/pgtable.h #define pte_dirty(pte)		(pte_sw_dirty(pte) || pte_hw_dirty(pte))
pte                98 arch/arm64/include/asm/pgtable.h #define pte_valid(pte)		(!!(pte_val(pte) & PTE_VALID))
pte                99 arch/arm64/include/asm/pgtable.h #define pte_valid_not_user(pte) \
pte               100 arch/arm64/include/asm/pgtable.h 	((pte_val(pte) & (PTE_VALID | PTE_USER)) == PTE_VALID)
pte               101 arch/arm64/include/asm/pgtable.h #define pte_valid_young(pte) \
pte               102 arch/arm64/include/asm/pgtable.h 	((pte_val(pte) & (PTE_VALID | PTE_AF)) == (PTE_VALID | PTE_AF))
pte               103 arch/arm64/include/asm/pgtable.h #define pte_valid_user(pte) \
pte               104 arch/arm64/include/asm/pgtable.h 	((pte_val(pte) & (PTE_VALID | PTE_USER)) == (PTE_VALID | PTE_USER))
pte               111 arch/arm64/include/asm/pgtable.h #define pte_accessible(mm, pte)	\
pte               112 arch/arm64/include/asm/pgtable.h 	(mm_tlb_flush_pending(mm) ? pte_present(pte) : pte_valid_young(pte))
pte               119 arch/arm64/include/asm/pgtable.h #define pte_access_permitted(pte, write) \
pte               120 arch/arm64/include/asm/pgtable.h 	(pte_valid_user(pte) && (!(write) || pte_write(pte)))
pte               126 arch/arm64/include/asm/pgtable.h static inline pte_t clear_pte_bit(pte_t pte, pgprot_t prot)
pte               128 arch/arm64/include/asm/pgtable.h 	pte_val(pte) &= ~pgprot_val(prot);
pte               129 arch/arm64/include/asm/pgtable.h 	return pte;
pte               132 arch/arm64/include/asm/pgtable.h static inline pte_t set_pte_bit(pte_t pte, pgprot_t prot)
pte               134 arch/arm64/include/asm/pgtable.h 	pte_val(pte) |= pgprot_val(prot);
pte               135 arch/arm64/include/asm/pgtable.h 	return pte;
pte               138 arch/arm64/include/asm/pgtable.h static inline pte_t pte_wrprotect(pte_t pte)
pte               140 arch/arm64/include/asm/pgtable.h 	pte = clear_pte_bit(pte, __pgprot(PTE_WRITE));
pte               141 arch/arm64/include/asm/pgtable.h 	pte = set_pte_bit(pte, __pgprot(PTE_RDONLY));
pte               142 arch/arm64/include/asm/pgtable.h 	return pte;
pte               145 arch/arm64/include/asm/pgtable.h static inline pte_t pte_mkwrite(pte_t pte)
pte               147 arch/arm64/include/asm/pgtable.h 	pte = set_pte_bit(pte, __pgprot(PTE_WRITE));
pte               148 arch/arm64/include/asm/pgtable.h 	pte = clear_pte_bit(pte, __pgprot(PTE_RDONLY));
pte               149 arch/arm64/include/asm/pgtable.h 	return pte;
pte               152 arch/arm64/include/asm/pgtable.h static inline pte_t pte_mkclean(pte_t pte)
pte               154 arch/arm64/include/asm/pgtable.h 	pte = clear_pte_bit(pte, __pgprot(PTE_DIRTY));
pte               155 arch/arm64/include/asm/pgtable.h 	pte = set_pte_bit(pte, __pgprot(PTE_RDONLY));
pte               157 arch/arm64/include/asm/pgtable.h 	return pte;
pte               160 arch/arm64/include/asm/pgtable.h static inline pte_t pte_mkdirty(pte_t pte)
pte               162 arch/arm64/include/asm/pgtable.h 	pte = set_pte_bit(pte, __pgprot(PTE_DIRTY));
pte               164 arch/arm64/include/asm/pgtable.h 	if (pte_write(pte))
pte               165 arch/arm64/include/asm/pgtable.h 		pte = clear_pte_bit(pte, __pgprot(PTE_RDONLY));
pte               167 arch/arm64/include/asm/pgtable.h 	return pte;
pte               170 arch/arm64/include/asm/pgtable.h static inline pte_t pte_mkold(pte_t pte)
pte               172 arch/arm64/include/asm/pgtable.h 	return clear_pte_bit(pte, __pgprot(PTE_AF));
pte               175 arch/arm64/include/asm/pgtable.h static inline pte_t pte_mkyoung(pte_t pte)
pte               177 arch/arm64/include/asm/pgtable.h 	return set_pte_bit(pte, __pgprot(PTE_AF));
pte               180 arch/arm64/include/asm/pgtable.h static inline pte_t pte_mkspecial(pte_t pte)
pte               182 arch/arm64/include/asm/pgtable.h 	return set_pte_bit(pte, __pgprot(PTE_SPECIAL));
pte               185 arch/arm64/include/asm/pgtable.h static inline pte_t pte_mkcont(pte_t pte)
pte               187 arch/arm64/include/asm/pgtable.h 	pte = set_pte_bit(pte, __pgprot(PTE_CONT));
pte               188 arch/arm64/include/asm/pgtable.h 	return set_pte_bit(pte, __pgprot(PTE_TYPE_PAGE));
pte               191 arch/arm64/include/asm/pgtable.h static inline pte_t pte_mknoncont(pte_t pte)
pte               193 arch/arm64/include/asm/pgtable.h 	return clear_pte_bit(pte, __pgprot(PTE_CONT));
pte               196 arch/arm64/include/asm/pgtable.h static inline pte_t pte_mkpresent(pte_t pte)
pte               198 arch/arm64/include/asm/pgtable.h 	return set_pte_bit(pte, __pgprot(PTE_VALID));
pte               206 arch/arm64/include/asm/pgtable.h static inline pte_t pte_mkdevmap(pte_t pte)
pte               208 arch/arm64/include/asm/pgtable.h 	return set_pte_bit(pte, __pgprot(PTE_DEVMAP | PTE_SPECIAL));
pte               211 arch/arm64/include/asm/pgtable.h static inline void set_pte(pte_t *ptep, pte_t pte)
pte               213 arch/arm64/include/asm/pgtable.h 	WRITE_ONCE(*ptep, pte);
pte               219 arch/arm64/include/asm/pgtable.h 	if (pte_valid_not_user(pte)) {
pte               244 arch/arm64/include/asm/pgtable.h 					   pte_t pte)
pte               253 arch/arm64/include/asm/pgtable.h 	if (!pte_valid(old_pte) || !pte_valid(pte))
pte               263 arch/arm64/include/asm/pgtable.h 	VM_WARN_ONCE(!pte_young(pte),
pte               265 arch/arm64/include/asm/pgtable.h 		     __func__, pte_val(old_pte), pte_val(pte));
pte               266 arch/arm64/include/asm/pgtable.h 	VM_WARN_ONCE(pte_write(old_pte) && !pte_dirty(pte),
pte               268 arch/arm64/include/asm/pgtable.h 		     __func__, pte_val(old_pte), pte_val(pte));
pte               272 arch/arm64/include/asm/pgtable.h 			      pte_t *ptep, pte_t pte)
pte               274 arch/arm64/include/asm/pgtable.h 	if (pte_present(pte) && pte_user_exec(pte) && !pte_special(pte))
pte               275 arch/arm64/include/asm/pgtable.h 		__sync_icache_dcache(pte);
pte               277 arch/arm64/include/asm/pgtable.h 	__check_racy_pte_update(mm, ptep, pte);
pte               279 arch/arm64/include/asm/pgtable.h 	set_pte(ptep, pte);
pte               285 arch/arm64/include/asm/pgtable.h #define pte_mkhuge(pte)		(__pte(pte_val(pte) & ~PTE_TABLE_BIT))
pte               306 arch/arm64/include/asm/pgtable.h static inline pud_t pte_pud(pte_t pte)
pte               308 arch/arm64/include/asm/pgtable.h 	return __pud(pte_val(pte));
pte               321 arch/arm64/include/asm/pgtable.h static inline pmd_t pte_pmd(pte_t pte)
pte               323 arch/arm64/include/asm/pgtable.h 	return __pmd(pte_val(pte));
pte               340 arch/arm64/include/asm/pgtable.h static inline int pte_protnone(pte_t pte)
pte               342 arch/arm64/include/asm/pgtable.h 	return (pte_val(pte) & (PTE_VALID | PTE_PROT_NONE)) == PTE_PROT_NONE;
pte               496 arch/arm64/include/asm/pgtable.h static inline void pte_unmap(pte_t *pte) { }
pte               658 arch/arm64/include/asm/pgtable.h static inline pte_t pte_modify(pte_t pte, pgprot_t newprot)
pte               663 arch/arm64/include/asm/pgtable.h 	if (pte_hw_dirty(pte))
pte               664 arch/arm64/include/asm/pgtable.h 		pte = pte_mkdirty(pte);
pte               665 arch/arm64/include/asm/pgtable.h 	pte_val(pte) = (pte_val(pte) & ~mask) | (pgprot_val(newprot) & mask);
pte               666 arch/arm64/include/asm/pgtable.h 	return pte;
pte               705 arch/arm64/include/asm/pgtable.h 	pte_t old_pte, pte;
pte               707 arch/arm64/include/asm/pgtable.h 	pte = READ_ONCE(*ptep);
pte               709 arch/arm64/include/asm/pgtable.h 		old_pte = pte;
pte               710 arch/arm64/include/asm/pgtable.h 		pte = pte_mkold(pte);
pte               711 arch/arm64/include/asm/pgtable.h 		pte_val(pte) = cmpxchg_relaxed(&pte_val(*ptep),
pte               712 arch/arm64/include/asm/pgtable.h 					       pte_val(old_pte), pte_val(pte));
pte               713 arch/arm64/include/asm/pgtable.h 	} while (pte_val(pte) != pte_val(old_pte));
pte               715 arch/arm64/include/asm/pgtable.h 	return pte_young(pte);
pte               779 arch/arm64/include/asm/pgtable.h 	pte_t old_pte, pte;
pte               781 arch/arm64/include/asm/pgtable.h 	pte = READ_ONCE(*ptep);
pte               783 arch/arm64/include/asm/pgtable.h 		old_pte = pte;
pte               788 arch/arm64/include/asm/pgtable.h 		if (pte_hw_dirty(pte))
pte               789 arch/arm64/include/asm/pgtable.h 			pte = pte_mkdirty(pte);
pte               790 arch/arm64/include/asm/pgtable.h 		pte = pte_wrprotect(pte);
pte               791 arch/arm64/include/asm/pgtable.h 		pte_val(pte) = cmpxchg_relaxed(&pte_val(*ptep),
pte               792 arch/arm64/include/asm/pgtable.h 					       pte_val(old_pte), pte_val(pte));
pte               793 arch/arm64/include/asm/pgtable.h 	} while (pte_val(pte) != pte_val(old_pte));
pte               830 arch/arm64/include/asm/pgtable.h #define __pte_to_swp_entry(pte)	((swp_entry_t) { pte_val(pte) })
pte                44 arch/arm64/include/asm/tlb.h static inline void __pte_free_tlb(struct mmu_gather *tlb, pgtable_t pte,
pte                47 arch/arm64/include/asm/tlb.h 	pgtable_pte_page_dtor(pte);
pte                48 arch/arm64/include/asm/tlb.h 	tlb_remove_table(tlb, pte);
pte                88 arch/arm64/kernel/efi.c 	pte_t pte = READ_ONCE(*ptep);
pte                91 arch/arm64/kernel/efi.c 		pte = set_pte_bit(pte, __pgprot(PTE_RDONLY));
pte                93 arch/arm64/kernel/efi.c 		pte = set_pte_bit(pte, __pgprot(PTE_PXN));
pte                94 arch/arm64/kernel/efi.c 	set_pte(ptep, pte);
pte               344 arch/arm64/kernel/hibernate.c 	pte_t pte = READ_ONCE(*src_ptep);
pte               346 arch/arm64/kernel/hibernate.c 	if (pte_valid(pte)) {
pte               352 arch/arm64/kernel/hibernate.c 		set_pte(dst_ptep, pte_mkwrite(pte));
pte               353 arch/arm64/kernel/hibernate.c 	} else if (debug_pagealloc_enabled() && !pte_none(pte)) {
pte               363 arch/arm64/kernel/hibernate.c 		BUG_ON(!pfn_valid(pte_pfn(pte)));
pte               365 arch/arm64/kernel/hibernate.c 		set_pte(dst_ptep, pte_mkpresent(pte_mkwrite(pte)));
pte               161 arch/arm64/mm/fault.c 		pte_t *ptep, pte;
pte               179 arch/arm64/mm/fault.c 		pte = READ_ONCE(*ptep);
pte               180 arch/arm64/mm/fault.c 		pr_cont(", pte=%016llx", pte_val(pte));
pte               202 arch/arm64/mm/fault.c 	pte_t pte = READ_ONCE(*ptep);
pte               204 arch/arm64/mm/fault.c 	if (pte_same(pte, entry))
pte               217 arch/arm64/mm/fault.c 	pteval = pte_val(pte);
pte                54 arch/arm64/mm/flush.c void __sync_icache_dcache(pte_t pte)
pte                56 arch/arm64/mm/flush.c 	struct page *page = pte_page(pte);
pte                59 arch/arm64/mm/hugetlbpage.c static inline pgprot_t pte_pgprot(pte_t pte)
pte                61 arch/arm64/mm/hugetlbpage.c 	unsigned long pfn = pte_pfn(pte);
pte                63 arch/arm64/mm/hugetlbpage.c 	return __pgprot(pte_val(pfn_pte(pfn, __pgprot(0))) ^ pte_val(pte));
pte               128 arch/arm64/mm/hugetlbpage.c 		pte_t pte = ptep_get_and_clear(mm, addr, ptep);
pte               135 arch/arm64/mm/hugetlbpage.c 		if (pte_dirty(pte))
pte               138 arch/arm64/mm/hugetlbpage.c 		if (pte_young(pte))
pte               174 arch/arm64/mm/hugetlbpage.c 			    pte_t *ptep, pte_t pte)
pte               186 arch/arm64/mm/hugetlbpage.c 	WARN_ON(!pte_present(pte));
pte               188 arch/arm64/mm/hugetlbpage.c 	if (!pte_cont(pte)) {
pte               189 arch/arm64/mm/hugetlbpage.c 		set_pte_at(mm, addr, ptep, pte);
pte               194 arch/arm64/mm/hugetlbpage.c 	pfn = pte_pfn(pte);
pte               196 arch/arm64/mm/hugetlbpage.c 	hugeprot = pte_pgprot(pte);
pte               205 arch/arm64/mm/hugetlbpage.c 			  pte_t *ptep, pte_t pte, unsigned long sz)
pte               213 arch/arm64/mm/hugetlbpage.c 		set_pte(ptep, pte);
pte               349 arch/arm64/mm/hugetlbpage.c static int __cont_access_flags_changed(pte_t *ptep, pte_t pte, int ncontig)
pte               353 arch/arm64/mm/hugetlbpage.c 	if (pte_write(pte) != pte_write(huge_ptep_get(ptep)))
pte               359 arch/arm64/mm/hugetlbpage.c 		if (pte_dirty(pte) != pte_dirty(orig_pte))
pte               362 arch/arm64/mm/hugetlbpage.c 		if (pte_young(pte) != pte_young(orig_pte))
pte               371 arch/arm64/mm/hugetlbpage.c 			       pte_t pte, int dirty)
pte               375 arch/arm64/mm/hugetlbpage.c 	unsigned long pfn = pte_pfn(pte), dpfn;
pte               379 arch/arm64/mm/hugetlbpage.c 	if (!pte_cont(pte))
pte               380 arch/arm64/mm/hugetlbpage.c 		return ptep_set_access_flags(vma, addr, ptep, pte, dirty);
pte               385 arch/arm64/mm/hugetlbpage.c 	if (!__cont_access_flags_changed(ptep, pte, ncontig))
pte               392 arch/arm64/mm/hugetlbpage.c 		pte = pte_mkdirty(pte);
pte               395 arch/arm64/mm/hugetlbpage.c 		pte = pte_mkyoung(pte);
pte               397 arch/arm64/mm/hugetlbpage.c 	hugeprot = pte_pgprot(pte);
pte               411 arch/arm64/mm/hugetlbpage.c 	pte_t pte;
pte               421 arch/arm64/mm/hugetlbpage.c 	pte = get_clear_flush(mm, addr, ptep, pgsize, ncontig);
pte               422 arch/arm64/mm/hugetlbpage.c 	pte = pte_wrprotect(pte);
pte               424 arch/arm64/mm/hugetlbpage.c 	hugeprot = pte_pgprot(pte);
pte               425 arch/arm64/mm/hugetlbpage.c 	pfn = pte_pfn(pte);
pte               696 arch/arm64/mm/mmu.c 	pte_t *ptep, pte;
pte               722 arch/arm64/mm/mmu.c 	pte = READ_ONCE(*ptep);
pte               723 arch/arm64/mm/mmu.c 	if (pte_none(pte))
pte               726 arch/arm64/mm/mmu.c 	return pfn_valid(pte_pfn(pte));
pte                25 arch/arm64/mm/pageattr.c 	pte_t pte = READ_ONCE(*ptep);
pte                27 arch/arm64/mm/pageattr.c 	pte = clear_pte_bit(pte, cdata->clear_mask);
pte                28 arch/arm64/mm/pageattr.c 	pte = set_pte_bit(pte, cdata->set_mask);
pte                30 arch/arm64/mm/pageattr.c 	set_pte(ptep, pte);
pte                47 arch/c6x/include/asm/pgtable.h #define __pte_to_swp_entry(pte)	((swp_entry_t) { pte_val(pte) })
pte                33 arch/csky/abiv2/cacheflush.c 		      pte_t *pte)
pte                38 arch/csky/abiv2/cacheflush.c 	pfn = pte_pfn(*pte);
pte                15 arch/csky/include/asm/pgalloc.h 					pte_t *pte)
pte                17 arch/csky/include/asm/pgalloc.h 	set_pmd(pmd, __pmd(__pa(pte)));
pte                21 arch/csky/include/asm/pgalloc.h 					pgtable_t pte)
pte                23 arch/csky/include/asm/pgalloc.h 	set_pmd(pmd, __pmd(__pa(page_address(pte))));
pte                32 arch/csky/include/asm/pgalloc.h 	pte_t *pte;
pte                35 arch/csky/include/asm/pgalloc.h 	pte = (pte_t *) __get_free_page(GFP_KERNEL);
pte                36 arch/csky/include/asm/pgalloc.h 	if (!pte)
pte                40 arch/csky/include/asm/pgalloc.h 		(pte + i)->pte_low = _PAGE_GLOBAL;
pte                42 arch/csky/include/asm/pgalloc.h 	return pte;
pte                72 arch/csky/include/asm/pgalloc.h #define __pte_free_tlb(tlb, pte, address)		\
pte                74 arch/csky/include/asm/pgalloc.h 	pgtable_pte_page_dtor(pte);			\
pte                75 arch/csky/include/asm/pgalloc.h 	tlb_remove_page(tlb, pte);			\
pte                49 arch/csky/include/asm/pgtable.h #define pte_none(pte)		(!(pte_val(pte) & ~_PAGE_GLOBAL))
pte                50 arch/csky/include/asm/pgtable.h #define pte_present(pte)	(pte_val(pte) & _PAGE_PRESENT)
pte                61 arch/csky/include/asm/pgtable.h #define pte_unmap(pte)	((void)(pte))
pte                67 arch/csky/include/asm/pgtable.h #define __pte_to_swp_entry(pte)		((swp_entry_t) { pte_val(pte) })
pte               113 arch/csky/include/asm/pgtable.h static inline int pte_special(pte_t pte) { return 0; }
pte               114 arch/csky/include/asm/pgtable.h static inline pte_t pte_mkspecial(pte_t pte) { return pte; }
pte               116 arch/csky/include/asm/pgtable.h static inline void set_pte(pte_t *p, pte_t pte)
pte               118 arch/csky/include/asm/pgtable.h 	*p = pte;
pte               173 arch/csky/include/asm/pgtable.h static inline int pte_read(pte_t pte)
pte               175 arch/csky/include/asm/pgtable.h 	return pte.pte_low & _PAGE_READ;
pte               178 arch/csky/include/asm/pgtable.h static inline int pte_write(pte_t pte)
pte               180 arch/csky/include/asm/pgtable.h 	return (pte).pte_low & _PAGE_WRITE;
pte               183 arch/csky/include/asm/pgtable.h static inline int pte_dirty(pte_t pte)
pte               185 arch/csky/include/asm/pgtable.h 	return (pte).pte_low & _PAGE_MODIFIED;
pte               188 arch/csky/include/asm/pgtable.h static inline int pte_young(pte_t pte)
pte               190 arch/csky/include/asm/pgtable.h 	return (pte).pte_low & _PAGE_ACCESSED;
pte               193 arch/csky/include/asm/pgtable.h static inline pte_t pte_wrprotect(pte_t pte)
pte               195 arch/csky/include/asm/pgtable.h 	pte_val(pte) &= ~(_PAGE_WRITE | _PAGE_DIRTY);
pte               196 arch/csky/include/asm/pgtable.h 	return pte;
pte               199 arch/csky/include/asm/pgtable.h static inline pte_t pte_mkclean(pte_t pte)
pte               201 arch/csky/include/asm/pgtable.h 	pte_val(pte) &= ~(_PAGE_MODIFIED|_PAGE_DIRTY);
pte               202 arch/csky/include/asm/pgtable.h 	return pte;
pte               205 arch/csky/include/asm/pgtable.h static inline pte_t pte_mkold(pte_t pte)
pte               207 arch/csky/include/asm/pgtable.h 	pte_val(pte) &= ~(_PAGE_ACCESSED|_PAGE_VALID);
pte               208 arch/csky/include/asm/pgtable.h 	return pte;
pte               211 arch/csky/include/asm/pgtable.h static inline pte_t pte_mkwrite(pte_t pte)
pte               213 arch/csky/include/asm/pgtable.h 	pte_val(pte) |= _PAGE_WRITE;
pte               214 arch/csky/include/asm/pgtable.h 	if (pte_val(pte) & _PAGE_MODIFIED)
pte               215 arch/csky/include/asm/pgtable.h 		pte_val(pte) |= _PAGE_DIRTY;
pte               216 arch/csky/include/asm/pgtable.h 	return pte;
pte               219 arch/csky/include/asm/pgtable.h static inline pte_t pte_mkdirty(pte_t pte)
pte               221 arch/csky/include/asm/pgtable.h 	pte_val(pte) |= _PAGE_MODIFIED;
pte               222 arch/csky/include/asm/pgtable.h 	if (pte_val(pte) & _PAGE_WRITE)
pte               223 arch/csky/include/asm/pgtable.h 		pte_val(pte) |= _PAGE_DIRTY;
pte               224 arch/csky/include/asm/pgtable.h 	return pte;
pte               227 arch/csky/include/asm/pgtable.h static inline pte_t pte_mkyoung(pte_t pte)
pte               229 arch/csky/include/asm/pgtable.h 	pte_val(pte) |= _PAGE_ACCESSED;
pte               230 arch/csky/include/asm/pgtable.h 	if (pte_val(pte) & _PAGE_READ)
pte               231 arch/csky/include/asm/pgtable.h 		pte_val(pte) |= _PAGE_VALID;
pte               232 arch/csky/include/asm/pgtable.h 	return pte;
pte               281 arch/csky/include/asm/pgtable.h static inline pte_t pte_modify(pte_t pte, pgprot_t newprot)
pte               283 arch/csky/include/asm/pgtable.h 	return __pte((pte_val(pte) & _PAGE_CHG_MASK) |
pte               304 arch/csky/include/asm/pgtable.h 		      pte_t *pte);
pte               110 arch/csky/mm/highmem.c 	pte_t *pte;
pte               116 arch/csky/mm/highmem.c 	pte = kmap_pte - (idx - FIX_KMAP_BEGIN);
pte               117 arch/csky/mm/highmem.c 	return pte_page(*pte);
pte               127 arch/csky/mm/highmem.c 	pte_t *pte;
pte               143 arch/csky/mm/highmem.c 					pte = (pte_t *) memblock_alloc_low(PAGE_SIZE, PAGE_SIZE);
pte               144 arch/csky/mm/highmem.c 					if (!pte)
pte               149 arch/csky/mm/highmem.c 					set_pmd(pmd, __pmd(__pa(pte)));
pte               150 arch/csky/mm/highmem.c 					BUG_ON(pte != pte_offset_kernel(pmd, 0));
pte               169 arch/csky/mm/highmem.c 	pte_t *pte;
pte               189 arch/csky/mm/highmem.c 	pte = pte_offset_kernel(pmd, vaddr);
pte               190 arch/csky/mm/highmem.c 	pkmap_page_table = pte;
pte                16 arch/h8300/include/asm/pgtable.h #define __pte_to_swp_entry(pte)	((swp_entry_t) { pte_val(pte) })
pte                22 arch/h8300/include/asm/pgtable.h static inline int pte_file(pte_t pte) { return 0; }
pte                69 arch/hexagon/include/asm/page.h typedef struct { unsigned long pte; } pte_t;
pte                74 arch/hexagon/include/asm/page.h #define pte_val(x)     ((x).pte)
pte                50 arch/hexagon/include/asm/pgalloc.h 				pgtable_t pte)
pte                56 arch/hexagon/include/asm/pgalloc.h 	set_pmd(pmd, __pmd(((unsigned long)page_to_pfn(pte) << PAGE_SHIFT) |
pte                70 arch/hexagon/include/asm/pgalloc.h 				       pte_t *pte)
pte                82 arch/hexagon/include/asm/pgalloc.h 	set_pmd(pmd, __pmd(((unsigned long)__pa(pte)) | HEXAGON_L1_PTE_SIZE));
pte                90 arch/hexagon/include/asm/pgalloc.h 	set_pmd(ppmd, __pmd(((unsigned long)__pa(pte)) | HEXAGON_L1_PTE_SIZE));
pte                95 arch/hexagon/include/asm/pgalloc.h #define __pte_free_tlb(tlb, pte, addr)		\
pte                97 arch/hexagon/include/asm/pgalloc.h 	pgtable_pte_page_dtor((pte));		\
pte                98 arch/hexagon/include/asm/pgalloc.h 	tlb_remove_page((tlb), (pte));		\
pte               161 arch/hexagon/include/asm/pgtable.h #define pte_special(pte)	0
pte               162 arch/hexagon/include/asm/pgtable.h #define pte_mkspecial(pte)	(pte)
pte               166 arch/hexagon/include/asm/pgtable.h #define pte_mkhuge(pte) __pte((pte_val(pte) & ~0x3) | HVM_HUGEPAGE_SIZE)
pte               173 arch/hexagon/include/asm/pgtable.h extern void sync_icache_dcache(pte_t pte);
pte               175 arch/hexagon/include/asm/pgtable.h #define pte_present_exec_user(pte) \
pte               176 arch/hexagon/include/asm/pgtable.h 	((pte_val(pte) & (_PAGE_EXECUTE | _PAGE_USER)) == \
pte               281 arch/hexagon/include/asm/pgtable.h static inline int pte_none(pte_t pte)
pte               283 arch/hexagon/include/asm/pgtable.h 	return pte_val(pte) == _NULL_PTE;
pte               289 arch/hexagon/include/asm/pgtable.h static inline int pte_present(pte_t pte)
pte               291 arch/hexagon/include/asm/pgtable.h 	return pte_val(pte) & _PAGE_PRESENT;
pte               301 arch/hexagon/include/asm/pgtable.h static inline pte_t pte_mkold(pte_t pte)
pte               303 arch/hexagon/include/asm/pgtable.h 	pte_val(pte) &= ~_PAGE_ACCESSED;
pte               304 arch/hexagon/include/asm/pgtable.h 	return pte;
pte               308 arch/hexagon/include/asm/pgtable.h static inline pte_t pte_mkyoung(pte_t pte)
pte               310 arch/hexagon/include/asm/pgtable.h 	pte_val(pte) |= _PAGE_ACCESSED;
pte               311 arch/hexagon/include/asm/pgtable.h 	return pte;
pte               315 arch/hexagon/include/asm/pgtable.h static inline pte_t pte_mkclean(pte_t pte)
pte               317 arch/hexagon/include/asm/pgtable.h 	pte_val(pte) &= ~_PAGE_DIRTY;
pte               318 arch/hexagon/include/asm/pgtable.h 	return pte;
pte               322 arch/hexagon/include/asm/pgtable.h static inline pte_t pte_mkdirty(pte_t pte)
pte               324 arch/hexagon/include/asm/pgtable.h 	pte_val(pte) |= _PAGE_DIRTY;
pte               325 arch/hexagon/include/asm/pgtable.h 	return pte;
pte               329 arch/hexagon/include/asm/pgtable.h static inline int pte_young(pte_t pte)
pte               331 arch/hexagon/include/asm/pgtable.h 	return pte_val(pte) & _PAGE_ACCESSED;
pte               335 arch/hexagon/include/asm/pgtable.h static inline int pte_dirty(pte_t pte)
pte               337 arch/hexagon/include/asm/pgtable.h 	return pte_val(pte) & _PAGE_DIRTY;
pte               341 arch/hexagon/include/asm/pgtable.h static inline pte_t pte_modify(pte_t pte, pgprot_t prot)
pte               343 arch/hexagon/include/asm/pgtable.h 	pte_val(pte) &= PAGE_MASK;
pte               344 arch/hexagon/include/asm/pgtable.h 	pte_val(pte) |= pgprot_val(prot);
pte               345 arch/hexagon/include/asm/pgtable.h 	return pte;
pte               349 arch/hexagon/include/asm/pgtable.h static inline pte_t pte_wrprotect(pte_t pte)
pte               351 arch/hexagon/include/asm/pgtable.h 	pte_val(pte) &= ~_PAGE_WRITE;
pte               352 arch/hexagon/include/asm/pgtable.h 	return pte;
pte               356 arch/hexagon/include/asm/pgtable.h static inline pte_t pte_mkwrite(pte_t pte)
pte               358 arch/hexagon/include/asm/pgtable.h 	pte_val(pte) |= _PAGE_WRITE;
pte               359 arch/hexagon/include/asm/pgtable.h 	return pte;
pte               363 arch/hexagon/include/asm/pgtable.h static inline pte_t pte_mkexec(pte_t pte)
pte               365 arch/hexagon/include/asm/pgtable.h 	pte_val(pte) |= _PAGE_EXECUTE;
pte               366 arch/hexagon/include/asm/pgtable.h 	return pte;
pte               370 arch/hexagon/include/asm/pgtable.h static inline int pte_read(pte_t pte)
pte               372 arch/hexagon/include/asm/pgtable.h 	return pte_val(pte) & _PAGE_READ;
pte               376 arch/hexagon/include/asm/pgtable.h static inline int pte_write(pte_t pte)
pte               378 arch/hexagon/include/asm/pgtable.h 	return pte_val(pte) & _PAGE_WRITE;
pte               383 arch/hexagon/include/asm/pgtable.h static inline int pte_exec(pte_t pte)
pte               385 arch/hexagon/include/asm/pgtable.h 	return pte_val(pte) & _PAGE_EXECUTE;
pte               389 arch/hexagon/include/asm/pgtable.h #define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val(pte) })
pte               398 arch/hexagon/include/asm/pgtable.h #define pte_pfn(pte) (pte_val(pte) >> PAGE_SHIFT)
pte               407 arch/hexagon/include/asm/pgtable.h #define set_pte_at(mm, addr, ptep, pte) set_pte(ptep, pte)
pte               412 arch/hexagon/include/asm/pgtable.h #define pte_unmap(pte)		do { } while (0)
pte               413 arch/hexagon/include/asm/pgtable.h #define pte_unmap_nested(pte)	do { } while (0)
pte                47 arch/hexagon/include/asm/vm_mmu.h #define __HVM_PTE_CVAL(pte)	(((pte) & __HVM_PTE_C) >> 6)
pte                74 arch/hexagon/mm/init.c void sync_icache_dcache(pte_t pte)
pte                79 arch/hexagon/mm/init.c 	page = pte_page(pte);
pte               176 arch/ia64/include/asm/page.h   typedef struct { unsigned long pte; } pte_t;
pte               185 arch/ia64/include/asm/page.h # define pte_val(x)	((x).pte)
pte                75 arch/ia64/include/asm/pgalloc.h pmd_populate(struct mm_struct *mm, pmd_t * pmd_entry, pgtable_t pte)
pte                77 arch/ia64/include/asm/pgalloc.h 	pmd_val(*pmd_entry) = page_to_phys(pte);
pte                82 arch/ia64/include/asm/pgalloc.h pmd_populate_kernel(struct mm_struct *mm, pmd_t * pmd_entry, pte_t * pte)
pte                84 arch/ia64/include/asm/pgalloc.h 	pmd_val(*pmd_entry) = __pa(pte);
pte                87 arch/ia64/include/asm/pgalloc.h #define __pte_free_tlb(tlb, pte, address)	pte_free((tlb)->mm, pte)
pte               265 arch/ia64/include/asm/pgtable.h #define pte_none(pte) 			(!pte_val(pte))
pte               266 arch/ia64/include/asm/pgtable.h #define pte_present(pte)		(pte_val(pte) & (_PAGE_P | _PAGE_PROTNONE))
pte               267 arch/ia64/include/asm/pgtable.h #define pte_clear(mm,addr,pte)		(pte_val(*(pte)) = 0UL)
pte               269 arch/ia64/include/asm/pgtable.h #define pte_page(pte)			virt_to_page(((pte_val(pte) & _PFN_MASK) + PAGE_OFFSET))
pte               297 arch/ia64/include/asm/pgtable.h #define pte_write(pte)	((unsigned) (((pte_val(pte) & _PAGE_AR_MASK) >> _PAGE_AR_SHIFT) - 2) <= 4)
pte               298 arch/ia64/include/asm/pgtable.h #define pte_exec(pte)		((pte_val(pte) & _PAGE_AR_RX) != 0)
pte               299 arch/ia64/include/asm/pgtable.h #define pte_dirty(pte)		((pte_val(pte) & _PAGE_D) != 0)
pte               300 arch/ia64/include/asm/pgtable.h #define pte_young(pte)		((pte_val(pte) & _PAGE_A) != 0)
pte               301 arch/ia64/include/asm/pgtable.h #define pte_special(pte)	0
pte               307 arch/ia64/include/asm/pgtable.h #define pte_wrprotect(pte)	(__pte(pte_val(pte) & ~_PAGE_AR_RW))
pte               308 arch/ia64/include/asm/pgtable.h #define pte_mkwrite(pte)	(__pte(pte_val(pte) | _PAGE_AR_RW))
pte               309 arch/ia64/include/asm/pgtable.h #define pte_mkold(pte)		(__pte(pte_val(pte) & ~_PAGE_A))
pte               310 arch/ia64/include/asm/pgtable.h #define pte_mkyoung(pte)	(__pte(pte_val(pte) | _PAGE_A))
pte               311 arch/ia64/include/asm/pgtable.h #define pte_mkclean(pte)	(__pte(pte_val(pte) & ~_PAGE_D))
pte               312 arch/ia64/include/asm/pgtable.h #define pte_mkdirty(pte)	(__pte(pte_val(pte) | _PAGE_D))
pte               313 arch/ia64/include/asm/pgtable.h #define pte_mkhuge(pte)		(__pte(pte_val(pte)))
pte               314 arch/ia64/include/asm/pgtable.h #define pte_mkspecial(pte)	(pte)
pte               325 arch/ia64/include/asm/pgtable.h #define pte_present_exec_user(pte)\
pte               326 arch/ia64/include/asm/pgtable.h 	((pte_val(pte) & (_PAGE_P | _PAGE_PL_MASK | _PAGE_AR_RX)) == \
pte               405 arch/ia64/include/asm/pgtable.h #define pte_unmap(pte)			do { } while (0)
pte               417 arch/ia64/include/asm/pgtable.h 	pte_t pte = *ptep;
pte               418 arch/ia64/include/asm/pgtable.h 	if (!pte_young(pte))
pte               420 arch/ia64/include/asm/pgtable.h 	set_pte_at(vma->vm_mm, addr, ptep, pte_mkold(pte));
pte               431 arch/ia64/include/asm/pgtable.h 	pte_t pte = *ptep;
pte               433 arch/ia64/include/asm/pgtable.h 	return pte;
pte               479 arch/ia64/include/asm/pgtable.h #define __pte_to_swp_entry(pte)		((swp_entry_t) { pte_val(pte) })
pte               481 arch/ia64/include/asm/processor.h 	  __u64 vmaddr, __u64 pte,
pte               488 arch/ia64/include/asm/processor.h 		ia64_itri(tr_num, pte);
pte               490 arch/ia64/include/asm/processor.h 		ia64_itrd(tr_num, pte);
pte               498 arch/ia64/include/asm/processor.h ia64_itc (__u64 target_mask, __u64 vmaddr, __u64 pte,
pte               506 arch/ia64/include/asm/processor.h 		ia64_itci(pte);
pte               508 arch/ia64/include/asm/processor.h 		ia64_itcd(pte);
pte                20 arch/ia64/include/asm/tlbflush.h 	u64 pte;
pte                24 arch/ia64/include/asm/tlbflush.h extern int ia64_itr_entry(u64 target_mask, u64 va, u64 pte, u64 log_size);
pte              1236 arch/ia64/kernel/mca.c 		if (p->pte & 0x1) {
pte              1245 arch/ia64/kernel/mca.c 				ia64_itr(0x1, i, p->ifa, p->pte, p->itir >> 2);
pte              1249 arch/ia64/kernel/mca.c 				ia64_itr(0x2, i, p->ifa, p->pte, p->itir >> 2);
pte                34 arch/ia64/mm/fault.c 	pte_t *ptep, pte;
pte                52 arch/ia64/mm/fault.c 	pte = *ptep;
pte                53 arch/ia64/mm/fault.c 	return pte_present(pte);
pte                35 arch/ia64/mm/hugetlbpage.c 	pte_t *pte = NULL;
pte                42 arch/ia64/mm/hugetlbpage.c 			pte = pte_alloc_map(mm, pmd, taddr);
pte                44 arch/ia64/mm/hugetlbpage.c 	return pte;
pte                54 arch/ia64/mm/hugetlbpage.c 	pte_t *pte = NULL;
pte                62 arch/ia64/mm/hugetlbpage.c 				pte = pte_offset_map(pmd, taddr);
pte                66 arch/ia64/mm/hugetlbpage.c 	return pte;
pte                56 arch/ia64/mm/init.c __ia64_sync_icache_dcache (pte_t pte)
pte                61 arch/ia64/mm/init.c 	page = pte_page(pte);
pte               213 arch/ia64/mm/init.c 	pte_t *pte;
pte               224 arch/ia64/mm/init.c 		pte = pte_alloc_kernel(pmd, address);
pte               225 arch/ia64/mm/init.c 		if (!pte)
pte               227 arch/ia64/mm/init.c 		if (!pte_none(*pte))
pte               229 arch/ia64/mm/init.c 		set_pte(pte, mk_pte(page, pgprot));
pte               387 arch/ia64/mm/init.c 		pte_t *pte;
pte               407 arch/ia64/mm/init.c 		pte = pte_offset_kernel(pmd, end_address);
pte               409 arch/ia64/mm/init.c 		if (pte_none(*pte)) {
pte               411 arch/ia64/mm/init.c 			pte++;
pte               435 arch/ia64/mm/init.c 	pte_t *pte;
pte               463 arch/ia64/mm/init.c 			pte = memblock_alloc_node(PAGE_SIZE, PAGE_SIZE, node);
pte               464 arch/ia64/mm/init.c 			if (!pte)
pte               466 arch/ia64/mm/init.c 			pmd_populate_kernel(&init_mm, pmd, pte);
pte               468 arch/ia64/mm/init.c 		pte = pte_offset_kernel(pmd, address);
pte               470 arch/ia64/mm/init.c 		if (pte_none(*pte)) {
pte               475 arch/ia64/mm/init.c 			set_pte(pte, pfn_pte(__pa(page) >> PAGE_SHIFT,
pte               455 arch/ia64/mm/tlb.c int ia64_itr_entry(u64 target_mask, u64 va, u64 pte, u64 log_size)
pte               475 arch/ia64/mm/tlb.c 			if (p->pte & 0x1)
pte               487 arch/ia64/mm/tlb.c 			if (p->pte & 0x1)
pte               499 arch/ia64/mm/tlb.c 			if (!((ia64_idtrs[cpu] + i)->pte & 0x1))
pte               503 arch/ia64/mm/tlb.c 			if (!((ia64_idtrs[cpu] + IA64_TR_ALLOC_MAX + i)->pte & 0x1))
pte               507 arch/ia64/mm/tlb.c 			if (!((ia64_idtrs[cpu] + i)->pte & 0x1) &&
pte               508 arch/ia64/mm/tlb.c 			    !((ia64_idtrs[cpu] + IA64_TR_ALLOC_MAX + i)->pte & 0x1))
pte               526 arch/ia64/mm/tlb.c 		ia64_itr(0x1, i, va, pte, log_size);
pte               530 arch/ia64/mm/tlb.c 		p->pte = pte;
pte               535 arch/ia64/mm/tlb.c 		ia64_itr(0x2, i, va, pte, log_size);
pte               539 arch/ia64/mm/tlb.c 		p->pte = pte;
pte               569 arch/ia64/mm/tlb.c 		if ((p->pte&0x1) && is_tr_overlap(p, p->ifa, p->itir>>2)) {
pte               570 arch/ia64/mm/tlb.c 			p->pte = 0;
pte               578 arch/ia64/mm/tlb.c 		if ((p->pte & 0x1) && is_tr_overlap(p, p->ifa, p->itir>>2)) {
pte               579 arch/ia64/mm/tlb.c 			p->pte = 0;
pte               586 arch/ia64/mm/tlb.c 		if (((ia64_idtrs[cpu] + i)->pte & 0x1) ||
pte               587 arch/ia64/mm/tlb.c 		    ((ia64_idtrs[cpu] + IA64_TR_ALLOC_MAX + i)->pte & 0x1))
pte                 8 arch/m68k/include/asm/mcf_pgalloc.h extern inline void pte_free_kernel(struct mm_struct *mm, pte_t *pte)
pte                10 arch/m68k/include/asm/mcf_pgalloc.h 	free_page((unsigned long) pte);
pte                37 arch/m68k/include/asm/mcf_pgalloc.h #define pmd_populate_kernel(mm, pmd, pte) (pmd_val(*pmd) = (unsigned long)(pte))
pte                53 arch/m68k/include/asm/mcf_pgalloc.h 	pte_t *pte;
pte                62 arch/m68k/include/asm/mcf_pgalloc.h 	pte = kmap(page);
pte                63 arch/m68k/include/asm/mcf_pgalloc.h 	if (pte) {
pte                64 arch/m68k/include/asm/mcf_pgalloc.h 		clear_page(pte);
pte                65 arch/m68k/include/asm/mcf_pgalloc.h 		__flush_page_to_ram(pte);
pte                66 arch/m68k/include/asm/mcf_pgalloc.h 		flush_tlb_kernel_page(pte);
pte                67 arch/m68k/include/asm/mcf_pgalloc.h 		nocache_page(pte);
pte               103 arch/m68k/include/asm/mcf_pgalloc.h #define pgd_populate(mm, pmd, pte) BUG()
pte               159 arch/m68k/include/asm/mcf_pgtable.h static inline pte_t pte_modify(pte_t pte, pgprot_t newprot)
pte               161 arch/m68k/include/asm/mcf_pgtable.h 	pte_val(pte) = (pte_val(pte) & CF_PAGE_CHG_MASK) | pgprot_val(newprot);
pte               162 arch/m68k/include/asm/mcf_pgtable.h 	return pte;
pte               172 arch/m68k/include/asm/mcf_pgtable.h #define __pte_page(pte)	((unsigned long) (pte_val(pte) & PAGE_MASK))
pte               175 arch/m68k/include/asm/mcf_pgtable.h static inline int pte_none(pte_t pte)
pte               177 arch/m68k/include/asm/mcf_pgtable.h 	return !pte_val(pte);
pte               180 arch/m68k/include/asm/mcf_pgtable.h static inline int pte_present(pte_t pte)
pte               182 arch/m68k/include/asm/mcf_pgtable.h 	return pte_val(pte) & CF_PAGE_VALID;
pte               191 arch/m68k/include/asm/mcf_pgtable.h #define pte_pagenr(pte)	((__pte_page(pte) - PAGE_OFFSET) >> PAGE_SHIFT)
pte               192 arch/m68k/include/asm/mcf_pgtable.h #define pte_page(pte)	virt_to_page(__pte_page(pte))
pte               221 arch/m68k/include/asm/mcf_pgtable.h static inline int pte_read(pte_t pte)
pte               223 arch/m68k/include/asm/mcf_pgtable.h 	return pte_val(pte) & CF_PAGE_READABLE;
pte               226 arch/m68k/include/asm/mcf_pgtable.h static inline int pte_write(pte_t pte)
pte               228 arch/m68k/include/asm/mcf_pgtable.h 	return pte_val(pte) & CF_PAGE_WRITABLE;
pte               231 arch/m68k/include/asm/mcf_pgtable.h static inline int pte_exec(pte_t pte)
pte               233 arch/m68k/include/asm/mcf_pgtable.h 	return pte_val(pte) & CF_PAGE_EXEC;
pte               236 arch/m68k/include/asm/mcf_pgtable.h static inline int pte_dirty(pte_t pte)
pte               238 arch/m68k/include/asm/mcf_pgtable.h 	return pte_val(pte) & CF_PAGE_DIRTY;
pte               241 arch/m68k/include/asm/mcf_pgtable.h static inline int pte_young(pte_t pte)
pte               243 arch/m68k/include/asm/mcf_pgtable.h 	return pte_val(pte) & CF_PAGE_ACCESSED;
pte               246 arch/m68k/include/asm/mcf_pgtable.h static inline int pte_special(pte_t pte)
pte               251 arch/m68k/include/asm/mcf_pgtable.h static inline pte_t pte_wrprotect(pte_t pte)
pte               253 arch/m68k/include/asm/mcf_pgtable.h 	pte_val(pte) &= ~CF_PAGE_WRITABLE;
pte               254 arch/m68k/include/asm/mcf_pgtable.h 	return pte;
pte               257 arch/m68k/include/asm/mcf_pgtable.h static inline pte_t pte_rdprotect(pte_t pte)
pte               259 arch/m68k/include/asm/mcf_pgtable.h 	pte_val(pte) &= ~CF_PAGE_READABLE;
pte               260 arch/m68k/include/asm/mcf_pgtable.h 	return pte;
pte               263 arch/m68k/include/asm/mcf_pgtable.h static inline pte_t pte_exprotect(pte_t pte)
pte               265 arch/m68k/include/asm/mcf_pgtable.h 	pte_val(pte) &= ~CF_PAGE_EXEC;
pte               266 arch/m68k/include/asm/mcf_pgtable.h 	return pte;
pte               269 arch/m68k/include/asm/mcf_pgtable.h static inline pte_t pte_mkclean(pte_t pte)
pte               271 arch/m68k/include/asm/mcf_pgtable.h 	pte_val(pte) &= ~CF_PAGE_DIRTY;
pte               272 arch/m68k/include/asm/mcf_pgtable.h 	return pte;
pte               275 arch/m68k/include/asm/mcf_pgtable.h static inline pte_t pte_mkold(pte_t pte)
pte               277 arch/m68k/include/asm/mcf_pgtable.h 	pte_val(pte) &= ~CF_PAGE_ACCESSED;
pte               278 arch/m68k/include/asm/mcf_pgtable.h 	return pte;
pte               281 arch/m68k/include/asm/mcf_pgtable.h static inline pte_t pte_mkwrite(pte_t pte)
pte               283 arch/m68k/include/asm/mcf_pgtable.h 	pte_val(pte) |= CF_PAGE_WRITABLE;
pte               284 arch/m68k/include/asm/mcf_pgtable.h 	return pte;
pte               287 arch/m68k/include/asm/mcf_pgtable.h static inline pte_t pte_mkread(pte_t pte)
pte               289 arch/m68k/include/asm/mcf_pgtable.h 	pte_val(pte) |= CF_PAGE_READABLE;
pte               290 arch/m68k/include/asm/mcf_pgtable.h 	return pte;
pte               293 arch/m68k/include/asm/mcf_pgtable.h static inline pte_t pte_mkexec(pte_t pte)
pte               295 arch/m68k/include/asm/mcf_pgtable.h 	pte_val(pte) |= CF_PAGE_EXEC;
pte               296 arch/m68k/include/asm/mcf_pgtable.h 	return pte;
pte               299 arch/m68k/include/asm/mcf_pgtable.h static inline pte_t pte_mkdirty(pte_t pte)
pte               301 arch/m68k/include/asm/mcf_pgtable.h 	pte_val(pte) |= CF_PAGE_DIRTY;
pte               302 arch/m68k/include/asm/mcf_pgtable.h 	return pte;
pte               305 arch/m68k/include/asm/mcf_pgtable.h static inline pte_t pte_mkyoung(pte_t pte)
pte               307 arch/m68k/include/asm/mcf_pgtable.h 	pte_val(pte) |= CF_PAGE_ACCESSED;
pte               308 arch/m68k/include/asm/mcf_pgtable.h 	return pte;
pte               311 arch/m68k/include/asm/mcf_pgtable.h static inline pte_t pte_mknocache(pte_t pte)
pte               313 arch/m68k/include/asm/mcf_pgtable.h 	pte_val(pte) |= 0x80 | (pte_val(pte) & ~0x40);
pte               314 arch/m68k/include/asm/mcf_pgtable.h 	return pte;
pte               317 arch/m68k/include/asm/mcf_pgtable.h static inline pte_t pte_mkcache(pte_t pte)
pte               319 arch/m68k/include/asm/mcf_pgtable.h 	pte_val(pte) &= ~CF_PAGE_NOCACHE;
pte               320 arch/m68k/include/asm/mcf_pgtable.h 	return pte;
pte               323 arch/m68k/include/asm/mcf_pgtable.h static inline pte_t pte_mkspecial(pte_t pte)
pte               325 arch/m68k/include/asm/mcf_pgtable.h 	return pte;
pte               396 arch/m68k/include/asm/mcf_pgtable.h #define __pte_to_swp_entry(pte)	((swp_entry_t) { pte_val(pte) })
pte               403 arch/m68k/include/asm/mcf_pgtable.h #define pte_unmap(pte)		((void) 0)
pte               405 arch/m68k/include/asm/mcf_pgtable.h #define pte_pfn(pte)		(pte_val(pte) >> PAGE_SHIFT)
pte               104 arch/m68k/include/asm/mmu_context.h 	pte_t *pte;
pte               134 arch/m68k/include/asm/mmu_context.h 	pte = (mmuar >= PAGE_OFFSET) ? pte_offset_kernel(pmd, mmuar)
pte               136 arch/m68k/include/asm/mmu_context.h 	if (pte_none(*pte) || !pte_present(*pte))
pte               139 arch/m68k/include/asm/mmu_context.h 	set_pte(pte, pte_mkyoung(*pte));
pte               141 arch/m68k/include/asm/mmu_context.h 	if (!pte_dirty(*pte) && mmuar <= PAGE_OFFSET)
pte               142 arch/m68k/include/asm/mmu_context.h 		set_pte(pte, pte_wrprotect(*pte));
pte               145 arch/m68k/include/asm/mmu_context.h 		(((int)(pte->pte) & (int)CF_PAGE_MMUTR_MASK)
pte               148 arch/m68k/include/asm/mmu_context.h 	mmu_write(MMUDR, (pte_val(*pte) & PAGE_MASK) |
pte               149 arch/m68k/include/asm/mmu_context.h 		((pte->pte) & CF_PAGE_MMUDR_MASK) | MMUDR_SZ_8KB | MMUDR_X);
pte                13 arch/m68k/include/asm/motorola_pgalloc.h 	pte_t *pte;
pte                15 arch/m68k/include/asm/motorola_pgalloc.h 	pte = (pte_t *)__get_free_page(GFP_KERNEL|__GFP_ZERO);
pte                16 arch/m68k/include/asm/motorola_pgalloc.h 	if (pte) {
pte                17 arch/m68k/include/asm/motorola_pgalloc.h 		__flush_page_to_ram(pte);
pte                18 arch/m68k/include/asm/motorola_pgalloc.h 		flush_tlb_kernel_page(pte);
pte                19 arch/m68k/include/asm/motorola_pgalloc.h 		nocache_page(pte);
pte                22 arch/m68k/include/asm/motorola_pgalloc.h 	return pte;
pte                25 arch/m68k/include/asm/motorola_pgalloc.h static inline void pte_free_kernel(struct mm_struct *mm, pte_t *pte)
pte                27 arch/m68k/include/asm/motorola_pgalloc.h 	cache_page(pte);
pte                28 arch/m68k/include/asm/motorola_pgalloc.h 	free_page((unsigned long) pte);
pte                34 arch/m68k/include/asm/motorola_pgalloc.h 	pte_t *pte;
pte                44 arch/m68k/include/asm/motorola_pgalloc.h 	pte = kmap(page);
pte                45 arch/m68k/include/asm/motorola_pgalloc.h 	__flush_page_to_ram(pte);
pte                46 arch/m68k/include/asm/motorola_pgalloc.h 	flush_tlb_kernel_page(pte);
pte                47 arch/m68k/include/asm/motorola_pgalloc.h 	nocache_page(pte);
pte                98 arch/m68k/include/asm/motorola_pgalloc.h static inline void pmd_populate_kernel(struct mm_struct *mm, pmd_t *pmd, pte_t *pte)
pte               100 arch/m68k/include/asm/motorola_pgalloc.h 	pmd_set(pmd, pte);
pte               103 arch/m68k/include/asm/motorola_pgtable.h static inline pte_t pte_modify(pte_t pte, pgprot_t newprot)
pte               105 arch/m68k/include/asm/motorola_pgtable.h 	pte_val(pte) = (pte_val(pte) & _PAGE_CHG_MASK) | pgprot_val(newprot);
pte               106 arch/m68k/include/asm/motorola_pgtable.h 	return pte;
pte               125 arch/m68k/include/asm/motorola_pgtable.h #define __pte_page(pte) ((unsigned long)__va(pte_val(pte) & PAGE_MASK))
pte               130 arch/m68k/include/asm/motorola_pgtable.h #define pte_none(pte)		(!pte_val(pte))
pte               131 arch/m68k/include/asm/motorola_pgtable.h #define pte_present(pte)	(pte_val(pte) & (_PAGE_PRESENT | _PAGE_PROTNONE))
pte               134 arch/m68k/include/asm/motorola_pgtable.h #define pte_page(pte)		virt_to_page(__va(pte_val(pte)))
pte               135 arch/m68k/include/asm/motorola_pgtable.h #define pte_pfn(pte)		(pte_val(pte) >> PAGE_SHIFT)
pte               168 arch/m68k/include/asm/motorola_pgtable.h static inline int pte_write(pte_t pte)		{ return !(pte_val(pte) & _PAGE_RONLY); }
pte               169 arch/m68k/include/asm/motorola_pgtable.h static inline int pte_dirty(pte_t pte)		{ return pte_val(pte) & _PAGE_DIRTY; }
pte               170 arch/m68k/include/asm/motorola_pgtable.h static inline int pte_young(pte_t pte)		{ return pte_val(pte) & _PAGE_ACCESSED; }
pte               171 arch/m68k/include/asm/motorola_pgtable.h static inline int pte_special(pte_t pte)	{ return 0; }
pte               173 arch/m68k/include/asm/motorola_pgtable.h static inline pte_t pte_wrprotect(pte_t pte)	{ pte_val(pte) |= _PAGE_RONLY; return pte; }
pte               174 arch/m68k/include/asm/motorola_pgtable.h static inline pte_t pte_mkclean(pte_t pte)	{ pte_val(pte) &= ~_PAGE_DIRTY; return pte; }
pte               175 arch/m68k/include/asm/motorola_pgtable.h static inline pte_t pte_mkold(pte_t pte)	{ pte_val(pte) &= ~_PAGE_ACCESSED; return pte; }
pte               176 arch/m68k/include/asm/motorola_pgtable.h static inline pte_t pte_mkwrite(pte_t pte)	{ pte_val(pte) &= ~_PAGE_RONLY; return pte; }
pte               177 arch/m68k/include/asm/motorola_pgtable.h static inline pte_t pte_mkdirty(pte_t pte)	{ pte_val(pte) |= _PAGE_DIRTY; return pte; }
pte               178 arch/m68k/include/asm/motorola_pgtable.h static inline pte_t pte_mkyoung(pte_t pte)	{ pte_val(pte) |= _PAGE_ACCESSED; return pte; }
pte               179 arch/m68k/include/asm/motorola_pgtable.h static inline pte_t pte_mknocache(pte_t pte)
pte               181 arch/m68k/include/asm/motorola_pgtable.h 	pte_val(pte) = (pte_val(pte) & _CACHEMASK040) | m68k_pgtable_cachemode;
pte               182 arch/m68k/include/asm/motorola_pgtable.h 	return pte;
pte               184 arch/m68k/include/asm/motorola_pgtable.h static inline pte_t pte_mkcache(pte_t pte)
pte               186 arch/m68k/include/asm/motorola_pgtable.h 	pte_val(pte) = (pte_val(pte) & _CACHEMASK040) | m68k_supervisor_cachemode;
pte               187 arch/m68k/include/asm/motorola_pgtable.h 	return pte;
pte               189 arch/m68k/include/asm/motorola_pgtable.h static inline pte_t pte_mkspecial(pte_t pte)	{ return pte; }
pte               224 arch/m68k/include/asm/motorola_pgtable.h #define pte_unmap(pte)		((void)0)
pte               272 arch/m68k/include/asm/motorola_pgtable.h #define __pte_to_swp_entry(pte)	((swp_entry_t) { pte_val(pte) })
pte                24 arch/m68k/include/asm/page.h typedef struct { unsigned long pte; } pte_t;
pte                30 arch/m68k/include/asm/page.h #define pte_val(x)	((x).pte)
pte                38 arch/m68k/include/asm/pgtable_no.h #define __pte_to_swp_entry(pte)	((swp_entry_t) { pte_val(pte) })
pte                22 arch/m68k/include/asm/sun3_pgalloc.h #define __pte_free_tlb(tlb,pte,addr)			\
pte                24 arch/m68k/include/asm/sun3_pgalloc.h 	pgtable_pte_page_dtor(pte);			\
pte                25 arch/m68k/include/asm/sun3_pgalloc.h 	tlb_remove_page((tlb), pte);			\
pte                28 arch/m68k/include/asm/sun3_pgalloc.h static inline void pmd_populate_kernel(struct mm_struct *mm, pmd_t *pmd, pte_t *pte)
pte                30 arch/m68k/include/asm/sun3_pgalloc.h 	pmd_val(*pmd) = __pa((unsigned long)pte);
pte                61 arch/m68k/include/asm/sun3_pgalloc.h #define pgd_populate(mm, pmd, pte) BUG()
pte               105 arch/m68k/include/asm/sun3_pgtable.h static inline pte_t pte_modify(pte_t pte, pgprot_t newprot)
pte               107 arch/m68k/include/asm/sun3_pgtable.h 	pte_val(pte) = (pte_val(pte) & SUN3_PAGE_CHG_MASK) | pgprot_val(newprot);
pte               108 arch/m68k/include/asm/sun3_pgtable.h 	return pte;
pte               118 arch/m68k/include/asm/sun3_pgtable.h #define __pte_page(pte) \
pte               119 arch/m68k/include/asm/sun3_pgtable.h ((unsigned long) __va ((pte_val (pte) & SUN3_PAGE_PGNUM_MASK) << PAGE_SHIFT))
pte               123 arch/m68k/include/asm/sun3_pgtable.h static inline int pte_none (pte_t pte) { return !pte_val (pte); }
pte               124 arch/m68k/include/asm/sun3_pgtable.h static inline int pte_present (pte_t pte) { return pte_val (pte) & SUN3_PAGE_VALID; }
pte               130 arch/m68k/include/asm/sun3_pgtable.h #define pte_pfn(pte)            (pte_val(pte) & SUN3_PAGE_PGNUM_MASK)
pte               134 arch/m68k/include/asm/sun3_pgtable.h #define pte_page(pte)		virt_to_page(__pte_page(pte))
pte               167 arch/m68k/include/asm/sun3_pgtable.h static inline int pte_write(pte_t pte)		{ return pte_val(pte) & SUN3_PAGE_WRITEABLE; }
pte               168 arch/m68k/include/asm/sun3_pgtable.h static inline int pte_dirty(pte_t pte)		{ return pte_val(pte) & SUN3_PAGE_MODIFIED; }
pte               169 arch/m68k/include/asm/sun3_pgtable.h static inline int pte_young(pte_t pte)		{ return pte_val(pte) & SUN3_PAGE_ACCESSED; }
pte               170 arch/m68k/include/asm/sun3_pgtable.h static inline int pte_special(pte_t pte)	{ return 0; }
pte               172 arch/m68k/include/asm/sun3_pgtable.h static inline pte_t pte_wrprotect(pte_t pte)	{ pte_val(pte) &= ~SUN3_PAGE_WRITEABLE; return pte; }
pte               173 arch/m68k/include/asm/sun3_pgtable.h static inline pte_t pte_mkclean(pte_t pte)	{ pte_val(pte) &= ~SUN3_PAGE_MODIFIED; return pte; }
pte               174 arch/m68k/include/asm/sun3_pgtable.h static inline pte_t pte_mkold(pte_t pte)	{ pte_val(pte) &= ~SUN3_PAGE_ACCESSED; return pte; }
pte               175 arch/m68k/include/asm/sun3_pgtable.h static inline pte_t pte_mkwrite(pte_t pte)	{ pte_val(pte) |= SUN3_PAGE_WRITEABLE; return pte; }
pte               176 arch/m68k/include/asm/sun3_pgtable.h static inline pte_t pte_mkdirty(pte_t pte)	{ pte_val(pte) |= SUN3_PAGE_MODIFIED; return pte; }
pte               177 arch/m68k/include/asm/sun3_pgtable.h static inline pte_t pte_mkyoung(pte_t pte)	{ pte_val(pte) |= SUN3_PAGE_ACCESSED; return pte; }
pte               178 arch/m68k/include/asm/sun3_pgtable.h static inline pte_t pte_mknocache(pte_t pte)	{ pte_val(pte) |= SUN3_PAGE_NOCACHE; return pte; }
pte               182 arch/m68k/include/asm/sun3_pgtable.h static inline pte_t pte_mkcache(pte_t pte)	{ return pte; }
pte               183 arch/m68k/include/asm/sun3_pgtable.h static inline pte_t pte_mkspecial(pte_t pte)	{ return pte; }
pte               207 arch/m68k/include/asm/sun3_pgtable.h #define pte_unmap(pte) do { } while (0)
pte               213 arch/m68k/include/asm/sun3_pgtable.h #define __pte_to_swp_entry(pte)	((swp_entry_t) { pte_val(pte) })
pte               469 arch/m68k/kernel/sys_m68k.c 		pte_t *pte;
pte               480 arch/m68k/kernel/sys_m68k.c 		pte = pte_offset_map_lock(mm, pmd, (unsigned long)mem, &ptl);
pte               481 arch/m68k/kernel/sys_m68k.c 		if (!pte_present(*pte) || !pte_dirty(*pte)
pte               482 arch/m68k/kernel/sys_m68k.c 		    || !pte_write(*pte)) {
pte               483 arch/m68k/kernel/sys_m68k.c 			pte_unmap_unlock(pte, ptl);
pte               495 arch/m68k/kernel/sys_m68k.c 		pte_unmap_unlock(pte, ptl);
pte                73 arch/m68k/mm/mcfmmu.c 			pte_t pte = pfn_pte(virt_to_pfn(address), PAGE_INIT);
pte                75 arch/m68k/mm/mcfmmu.c 				pte_val(pte) = 0;
pte                77 arch/m68k/mm/mcfmmu.c 			set_pte(pg_table, pte);
pte                96 arch/m68k/mm/mcfmmu.c 	pte_t *pte;
pte               122 arch/m68k/mm/mcfmmu.c 	pte = (KMAPAREA(mmuar)) ? pte_offset_kernel(pmd, mmuar)
pte               124 arch/m68k/mm/mcfmmu.c 	if (pte_none(*pte) || !pte_present(*pte)) {
pte               130 arch/m68k/mm/mcfmmu.c 		if (!pte_write(*pte)) {
pte               134 arch/m68k/mm/mcfmmu.c 		set_pte(pte, pte_mkdirty(*pte));
pte               137 arch/m68k/mm/mcfmmu.c 	set_pte(pte, pte_mkyoung(*pte));
pte               139 arch/m68k/mm/mcfmmu.c 	if (!pte_dirty(*pte) && !KMAPAREA(mmuar))
pte               140 arch/m68k/mm/mcfmmu.c 		set_pte(pte, pte_wrprotect(*pte));
pte               144 arch/m68k/mm/mcfmmu.c 		mmutr |= (pte->pte & CF_PAGE_MMUTR_MASK) >> CF_PAGE_MMUTR_SHIFT;
pte               147 arch/m68k/mm/mcfmmu.c 	mmu_write(MMUDR, (pte_val(*pte) & PAGE_MASK) |
pte               148 arch/m68k/mm/mcfmmu.c 		((pte->pte) & CF_PAGE_MMUDR_MASK) | MMUDR_SZ_8KB | MMUDR_X);
pte                33 arch/m68k/mm/sun3kmap.c 	unsigned long pte;
pte                37 arch/m68k/mm/sun3kmap.c 	pte = pte_val(ptep);
pte                38 arch/m68k/mm/sun3kmap.c 	pte |= type;
pte                40 arch/m68k/mm/sun3kmap.c 	sun3_put_pte(virt, pte);
pte                79 arch/m68k/mm/sun3mmu.c 			pte_t pte = pfn_pte(virt_to_pfn(address), PAGE_INIT);
pte                81 arch/m68k/mm/sun3mmu.c 				pte_val (pte) = 0;
pte                82 arch/m68k/mm/sun3mmu.c 			set_pte (pg_table, pte);
pte                26 arch/m68k/sun3/dvma.c 	unsigned long pte;
pte                34 arch/m68k/sun3/dvma.c 	pte = pte_val(ptep);
pte                36 arch/m68k/sun3/dvma.c 	if(ptelist[(vaddr & 0xff000) >> PAGE_SHIFT] != pte) {
pte                37 arch/m68k/sun3/dvma.c 		sun3_put_pte(vaddr, pte);
pte                38 arch/m68k/sun3/dvma.c 		ptelist[(vaddr & 0xff000) >> PAGE_SHIFT] = pte;
pte                72 arch/m68k/sun3/mmu_emu.c void print_pte (pte_t pte)
pte                76 arch/m68k/sun3/mmu_emu.c 	unsigned long val = pte_val (pte);
pte                94 arch/m68k/sun3/mmu_emu.c 	unsigned long val = pte_val (pte);
pte               360 arch/m68k/sun3/mmu_emu.c 	pte_t *pte;
pte               387 arch/m68k/sun3/mmu_emu.c 	pte = (pte_t *) pgd_val (*(crp + segment));
pte               390 arch/m68k/sun3/mmu_emu.c 	if (!pte) {
pte               395 arch/m68k/sun3/mmu_emu.c 	pte = (pte_t *) __va ((unsigned long)(pte + offset));
pte               398 arch/m68k/sun3/mmu_emu.c 	if (!(pte_val (*pte) & SUN3_PAGE_VALID))
pte               406 arch/m68k/sun3/mmu_emu.c 	sun3_put_pte (vaddr&PAGE_MASK, pte_val (*pte));
pte               414 arch/m68k/sun3/mmu_emu.c 		if (pte_val (*pte) & SUN3_PAGE_WRITEABLE)
pte               415 arch/m68k/sun3/mmu_emu.c 			pte_val (*pte) |= (SUN3_PAGE_ACCESSED
pte               420 arch/m68k/sun3/mmu_emu.c 		pte_val (*pte) |= SUN3_PAGE_ACCESSED;
pte               109 arch/m68k/sun3x/dvma.c 			pte_t *pte;
pte               112 arch/m68k/sun3x/dvma.c 			if((pte = pte_alloc_kernel(pmd, vaddr)) == NULL) {
pte               125 arch/m68k/sun3x/dvma.c 				set_pte(pte, pfn_pte(virt_to_pfn(kaddr),
pte               127 arch/m68k/sun3x/dvma.c 				pte++;
pte                89 arch/microblaze/include/asm/page.h typedef struct { unsigned long	pte; }		pte_t;
pte               102 arch/microblaze/include/asm/page.h # define pte_val(x)	((x).pte)
pte                53 arch/microblaze/include/asm/pgalloc.h #define __pte_free_tlb(tlb, pte, addr)	pte_free((tlb)->mm, (pte))
pte                55 arch/microblaze/include/asm/pgalloc.h #define pmd_populate(mm, pmd, pte) \
pte                56 arch/microblaze/include/asm/pgalloc.h 			(pmd_val(*(pmd)) = (unsigned long)page_address(pte))
pte                58 arch/microblaze/include/asm/pgalloc.h #define pmd_populate_kernel(mm, pmd, pte) \
pte                59 arch/microblaze/include/asm/pgalloc.h 		(pmd_val(*(pmd)) = (unsigned long) (pte))
pte                68 arch/microblaze/include/asm/pgalloc.h #define pgd_populate(mm, pmd, pte)	BUG()
pte                42 arch/microblaze/include/asm/pgtable.h #define __pte_to_swp_entry(pte)	((swp_entry_t) { pte_val(pte) })
pte                85 arch/microblaze/include/asm/pgtable.h static inline int pte_special(pte_t pte)	{ return 0; }
pte                87 arch/microblaze/include/asm/pgtable.h static inline pte_t pte_mkspecial(pte_t pte)	{ return pte; }
pte               296 arch/microblaze/include/asm/pgtable.h #define pte_none(pte)		((pte_val(pte) & ~_PTE_NONE_MASK) == 0)
pte               297 arch/microblaze/include/asm/pgtable.h #define pte_present(pte)	(pte_val(pte) & _PAGE_PRESENT)
pte               332 arch/microblaze/include/asm/pgtable.h static inline int pte_read(pte_t pte)  { return pte_val(pte) & _PAGE_USER; }
pte               333 arch/microblaze/include/asm/pgtable.h static inline int pte_write(pte_t pte) { return pte_val(pte) & _PAGE_RW; }
pte               334 arch/microblaze/include/asm/pgtable.h static inline int pte_exec(pte_t pte)  { return pte_val(pte) & _PAGE_EXEC; }
pte               335 arch/microblaze/include/asm/pgtable.h static inline int pte_dirty(pte_t pte) { return pte_val(pte) & _PAGE_DIRTY; }
pte               336 arch/microblaze/include/asm/pgtable.h static inline int pte_young(pte_t pte) { return pte_val(pte) & _PAGE_ACCESSED; }
pte               338 arch/microblaze/include/asm/pgtable.h static inline void pte_uncache(pte_t pte) { pte_val(pte) |= _PAGE_NO_CACHE; }
pte               339 arch/microblaze/include/asm/pgtable.h static inline void pte_cache(pte_t pte)   { pte_val(pte) &= ~_PAGE_NO_CACHE; }
pte               341 arch/microblaze/include/asm/pgtable.h static inline pte_t pte_rdprotect(pte_t pte) \
pte               342 arch/microblaze/include/asm/pgtable.h 		{ pte_val(pte) &= ~_PAGE_USER; return pte; }
pte               343 arch/microblaze/include/asm/pgtable.h static inline pte_t pte_wrprotect(pte_t pte) \
pte               344 arch/microblaze/include/asm/pgtable.h 	{ pte_val(pte) &= ~(_PAGE_RW | _PAGE_HWWRITE); return pte; }
pte               345 arch/microblaze/include/asm/pgtable.h static inline pte_t pte_exprotect(pte_t pte) \
pte               346 arch/microblaze/include/asm/pgtable.h 	{ pte_val(pte) &= ~_PAGE_EXEC; return pte; }
pte               347 arch/microblaze/include/asm/pgtable.h static inline pte_t pte_mkclean(pte_t pte) \
pte               348 arch/microblaze/include/asm/pgtable.h 	{ pte_val(pte) &= ~(_PAGE_DIRTY | _PAGE_HWWRITE); return pte; }
pte               349 arch/microblaze/include/asm/pgtable.h static inline pte_t pte_mkold(pte_t pte) \
pte               350 arch/microblaze/include/asm/pgtable.h 	{ pte_val(pte) &= ~_PAGE_ACCESSED; return pte; }
pte               352 arch/microblaze/include/asm/pgtable.h static inline pte_t pte_mkread(pte_t pte) \
pte               353 arch/microblaze/include/asm/pgtable.h 	{ pte_val(pte) |= _PAGE_USER; return pte; }
pte               354 arch/microblaze/include/asm/pgtable.h static inline pte_t pte_mkexec(pte_t pte) \
pte               355 arch/microblaze/include/asm/pgtable.h 	{ pte_val(pte) |= _PAGE_USER | _PAGE_EXEC; return pte; }
pte               356 arch/microblaze/include/asm/pgtable.h static inline pte_t pte_mkwrite(pte_t pte) \
pte               357 arch/microblaze/include/asm/pgtable.h 	{ pte_val(pte) |= _PAGE_RW; return pte; }
pte               358 arch/microblaze/include/asm/pgtable.h static inline pte_t pte_mkdirty(pte_t pte) \
pte               359 arch/microblaze/include/asm/pgtable.h 	{ pte_val(pte) |= _PAGE_DIRTY; return pte; }
pte               360 arch/microblaze/include/asm/pgtable.h static inline pte_t pte_mkyoung(pte_t pte) \
pte               361 arch/microblaze/include/asm/pgtable.h 	{ pte_val(pte) |= _PAGE_ACCESSED; return pte; }
pte               370 arch/microblaze/include/asm/pgtable.h 	pte_t pte;
pte               371 arch/microblaze/include/asm/pgtable.h 	pte_val(pte) = physpage | pgprot_val(pgprot);
pte               372 arch/microblaze/include/asm/pgtable.h 	return pte;
pte               377 arch/microblaze/include/asm/pgtable.h 	pte_t pte;							   \
pte               378 arch/microblaze/include/asm/pgtable.h 	pte_val(pte) = (((page - mem_map) << PAGE_SHIFT) + memory_start) |  \
pte               380 arch/microblaze/include/asm/pgtable.h 	pte;								   \
pte               383 arch/microblaze/include/asm/pgtable.h static inline pte_t pte_modify(pte_t pte, pgprot_t newprot)
pte               385 arch/microblaze/include/asm/pgtable.h 	pte_val(pte) = (pte_val(pte) & _PAGE_CHG_MASK) | pgprot_val(newprot);
pte               386 arch/microblaze/include/asm/pgtable.h 	return pte;
pte               421 arch/microblaze/include/asm/pgtable.h 		pte_t *ptep, pte_t pte)
pte               423 arch/microblaze/include/asm/pgtable.h 	*ptep = pte;
pte               427 arch/microblaze/include/asm/pgtable.h 		pte_t *ptep, pte_t pte)
pte               429 arch/microblaze/include/asm/pgtable.h 	*ptep = pte;
pte               496 arch/microblaze/include/asm/pgtable.h #define pte_unmap(pte)		kunmap_atomic(pte)
pte               510 arch/microblaze/include/asm/pgtable.h #define __pte_to_swp_entry(pte)	((swp_entry_t) { pte_val(pte) >> 2 })
pte               192 arch/microblaze/mm/pgtable.c 	pte_t	*pte;
pte               199 arch/microblaze/mm/pgtable.c 			pte = pte_offset_kernel(pmd, addr & PAGE_MASK);
pte               200 arch/microblaze/mm/pgtable.c 			if (pte) {
pte               202 arch/microblaze/mm/pgtable.c 				*ptep = pte;
pte               216 arch/microblaze/mm/pgtable.c 	pte_t *pte;
pte               228 arch/microblaze/mm/pgtable.c 	if (get_pteptr(mm, addr, &pte))
pte               229 arch/microblaze/mm/pgtable.c 		pa = (pte_val(*pte) & PAGE_MASK) | (addr & ~PAGE_MASK);
pte               236 arch/microblaze/mm/pgtable.c 	pte_t *pte;
pte               238 arch/microblaze/mm/pgtable.c 		pte = (pte_t *)__get_free_page(GFP_KERNEL | __GFP_ZERO);
pte               240 arch/microblaze/mm/pgtable.c 		pte = (pte_t *)early_get_page();
pte               241 arch/microblaze/mm/pgtable.c 		if (pte)
pte               242 arch/microblaze/mm/pgtable.c 			clear_page(pte);
pte               244 arch/microblaze/mm/pgtable.c 	return pte;
pte                45 arch/mips/include/asm/hugetlb.h 	pte_t pte = *ptep;
pte                49 arch/mips/include/asm/hugetlb.h 	return pte;
pte                60 arch/mips/include/asm/hugetlb.h static inline int huge_pte_none(pte_t pte)
pte                62 arch/mips/include/asm/hugetlb.h 	unsigned long val = pte_val(pte) & ~_PAGE_GLOBAL;
pte                69 arch/mips/include/asm/hugetlb.h 					     pte_t *ptep, pte_t pte,
pte                72 arch/mips/include/asm/hugetlb.h 	int changed = !pte_same(*ptep, pte);
pte                75 arch/mips/include/asm/hugetlb.h 		set_pte_at(vma->vm_mm, addr, ptep, pte);
pte               943 arch/mips/include/asm/kvm_host.h int kvm_set_spte_hva(struct kvm *kvm, unsigned long hva, pte_t pte);
pte               128 arch/mips/include/asm/page.h      typedef struct { unsigned long long pte; } pte_t;
pte               129 arch/mips/include/asm/page.h      #define pte_val(x) ((x).pte)
pte               133 arch/mips/include/asm/page.h typedef struct { unsigned long pte; } pte_t;
pte               134 arch/mips/include/asm/page.h #define pte_val(x)	((x).pte)
pte                19 arch/mips/include/asm/pgalloc.h 	pte_t *pte)
pte                21 arch/mips/include/asm/pgalloc.h 	set_pmd(pmd, __pmd((unsigned long)pte));
pte                25 arch/mips/include/asm/pgalloc.h 	pgtable_t pte)
pte                27 arch/mips/include/asm/pgalloc.h 	set_pmd(pmd, __pmd((unsigned long)page_address(pte)));
pte                55 arch/mips/include/asm/pgalloc.h #define __pte_free_tlb(tlb,pte,address)			\
pte                57 arch/mips/include/asm/pgalloc.h 	pgtable_pte_page_dtor(pte);			\
pte                58 arch/mips/include/asm/pgalloc.h 	tlb_remove_page((tlb), pte);			\
pte               162 arch/mips/include/asm/pgtable-32.h 	pte_t pte;
pte               164 arch/mips/include/asm/pgtable-32.h 	pte.pte_low = (pfn >> _PAGE_PRESENT_SHIFT) |
pte               166 arch/mips/include/asm/pgtable-32.h 	pte.pte_high = (pfn << _PFN_SHIFT) |
pte               168 arch/mips/include/asm/pgtable-32.h 	return pte;
pte               177 arch/mips/include/asm/pgtable-32.h 	pte_t pte;
pte               179 arch/mips/include/asm/pgtable-32.h 	pte.pte_high = (pfn << 6) | (pgprot_val(prot) & 0x3f);
pte               180 arch/mips/include/asm/pgtable-32.h 	pte.pte_low = pgprot_val(prot);
pte               182 arch/mips/include/asm/pgtable-32.h 	return pte;
pte               188 arch/mips/include/asm/pgtable-32.h #define pte_pfn(x)		((unsigned long)((x).pte >> (PAGE_SHIFT + 2)))
pte               191 arch/mips/include/asm/pgtable-32.h #define pte_pfn(x)		((unsigned long)((x).pte >> _PFN_SHIFT))
pte               222 arch/mips/include/asm/pgtable-32.h #define pte_unmap(pte) ((void)(pte))
pte               230 arch/mips/include/asm/pgtable-32.h #define __pte_to_swp_entry(pte)		((swp_entry_t) { pte_val(pte) })
pte               241 arch/mips/include/asm/pgtable-32.h #define __pte_to_swp_entry(pte)		((swp_entry_t) { (pte).pte_high })
pte               250 arch/mips/include/asm/pgtable-32.h #define __pte_to_swp_entry(pte)		((swp_entry_t) { (pte).pte_high })
pte               264 arch/mips/include/asm/pgtable-32.h #define __pte_to_swp_entry(pte)		((swp_entry_t) { pte_val(pte) })
pte               314 arch/mips/include/asm/pgtable-64.h #define pte_pfn(x)		((unsigned long)((x).pte >> (PAGE_SHIFT + 2)))
pte               317 arch/mips/include/asm/pgtable-64.h #define pte_pfn(x)		((unsigned long)((x).pte >> _PFN_SHIFT))
pte               359 arch/mips/include/asm/pgtable-64.h #define pte_unmap(pte) ((void)(pte))
pte               373 arch/mips/include/asm/pgtable-64.h { pte_t pte; pte_val(pte) = (type << 16) | (offset << 24); return pte; }
pte               378 arch/mips/include/asm/pgtable-64.h #define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val(pte) })
pte               139 arch/mips/include/asm/pgtable.h # define pte_none(pte)		(!(((pte).pte_high) & ~_PAGE_GLOBAL))
pte               141 arch/mips/include/asm/pgtable.h # define pte_none(pte)		(!(((pte).pte_low | (pte).pte_high) & ~_PAGE_GLOBAL))
pte               144 arch/mips/include/asm/pgtable.h #define pte_present(pte)	((pte).pte_low & _PAGE_PRESENT)
pte               145 arch/mips/include/asm/pgtable.h #define pte_no_exec(pte)	((pte).pte_low & _PAGE_NO_EXEC)
pte               147 arch/mips/include/asm/pgtable.h static inline void set_pte(pte_t *ptep, pte_t pte)
pte               149 arch/mips/include/asm/pgtable.h 	ptep->pte_high = pte.pte_high;
pte               151 arch/mips/include/asm/pgtable.h 	ptep->pte_low = pte.pte_low;
pte               154 arch/mips/include/asm/pgtable.h 	if (pte.pte_high & _PAGE_GLOBAL) {
pte               156 arch/mips/include/asm/pgtable.h 	if (pte.pte_low & _PAGE_GLOBAL) {
pte               190 arch/mips/include/asm/pgtable.h #define pte_none(pte)		(!(pte_val(pte) & ~_PAGE_GLOBAL))
pte               191 arch/mips/include/asm/pgtable.h #define pte_present(pte)	(pte_val(pte) & _PAGE_PRESENT)
pte               192 arch/mips/include/asm/pgtable.h #define pte_no_exec(pte)	(pte_val(pte) & _PAGE_NO_EXEC)
pte               210 arch/mips/include/asm/pgtable.h 		cmpxchg64(&buddy->pte, 0, _PAGE_GLOBAL);
pte               212 arch/mips/include/asm/pgtable.h 		cmpxchg(&buddy->pte, 0, _PAGE_GLOBAL);
pte               235 arch/mips/include/asm/pgtable.h 	extern void __update_cache(unsigned long address, pte_t pte);
pte               277 arch/mips/include/asm/pgtable.h static inline int pte_write(pte_t pte)	{ return pte.pte_low & _PAGE_WRITE; }
pte               278 arch/mips/include/asm/pgtable.h static inline int pte_dirty(pte_t pte)	{ return pte.pte_low & _PAGE_MODIFIED; }
pte               279 arch/mips/include/asm/pgtable.h static inline int pte_young(pte_t pte)	{ return pte.pte_low & _PAGE_ACCESSED; }
pte               280 arch/mips/include/asm/pgtable.h static inline int pte_special(pte_t pte) { return pte.pte_low & _PAGE_SPECIAL; }
pte               282 arch/mips/include/asm/pgtable.h static inline pte_t pte_wrprotect(pte_t pte)
pte               284 arch/mips/include/asm/pgtable.h 	pte.pte_low  &= ~_PAGE_WRITE;
pte               286 arch/mips/include/asm/pgtable.h 		pte.pte_low &= ~_PAGE_SILENT_WRITE;
pte               287 arch/mips/include/asm/pgtable.h 	pte.pte_high &= ~_PAGE_SILENT_WRITE;
pte               288 arch/mips/include/asm/pgtable.h 	return pte;
pte               291 arch/mips/include/asm/pgtable.h static inline pte_t pte_mkclean(pte_t pte)
pte               293 arch/mips/include/asm/pgtable.h 	pte.pte_low  &= ~_PAGE_MODIFIED;
pte               295 arch/mips/include/asm/pgtable.h 		pte.pte_low &= ~_PAGE_SILENT_WRITE;
pte               296 arch/mips/include/asm/pgtable.h 	pte.pte_high &= ~_PAGE_SILENT_WRITE;
pte               297 arch/mips/include/asm/pgtable.h 	return pte;
pte               300 arch/mips/include/asm/pgtable.h static inline pte_t pte_mkold(pte_t pte)
pte               302 arch/mips/include/asm/pgtable.h 	pte.pte_low  &= ~_PAGE_ACCESSED;
pte               304 arch/mips/include/asm/pgtable.h 		pte.pte_low &= ~_PAGE_SILENT_READ;
pte               305 arch/mips/include/asm/pgtable.h 	pte.pte_high &= ~_PAGE_SILENT_READ;
pte               306 arch/mips/include/asm/pgtable.h 	return pte;
pte               309 arch/mips/include/asm/pgtable.h static inline pte_t pte_mkwrite(pte_t pte)
pte               311 arch/mips/include/asm/pgtable.h 	pte.pte_low |= _PAGE_WRITE;
pte               312 arch/mips/include/asm/pgtable.h 	if (pte.pte_low & _PAGE_MODIFIED) {
pte               314 arch/mips/include/asm/pgtable.h 			pte.pte_low |= _PAGE_SILENT_WRITE;
pte               315 arch/mips/include/asm/pgtable.h 		pte.pte_high |= _PAGE_SILENT_WRITE;
pte               317 arch/mips/include/asm/pgtable.h 	return pte;
pte               320 arch/mips/include/asm/pgtable.h static inline pte_t pte_mkdirty(pte_t pte)
pte               322 arch/mips/include/asm/pgtable.h 	pte.pte_low |= _PAGE_MODIFIED;
pte               323 arch/mips/include/asm/pgtable.h 	if (pte.pte_low & _PAGE_WRITE) {
pte               325 arch/mips/include/asm/pgtable.h 			pte.pte_low |= _PAGE_SILENT_WRITE;
pte               326 arch/mips/include/asm/pgtable.h 		pte.pte_high |= _PAGE_SILENT_WRITE;
pte               328 arch/mips/include/asm/pgtable.h 	return pte;
pte               331 arch/mips/include/asm/pgtable.h static inline pte_t pte_mkyoung(pte_t pte)
pte               333 arch/mips/include/asm/pgtable.h 	pte.pte_low |= _PAGE_ACCESSED;
pte               334 arch/mips/include/asm/pgtable.h 	if (!(pte.pte_low & _PAGE_NO_READ)) {
pte               336 arch/mips/include/asm/pgtable.h 			pte.pte_low |= _PAGE_SILENT_READ;
pte               337 arch/mips/include/asm/pgtable.h 		pte.pte_high |= _PAGE_SILENT_READ;
pte               339 arch/mips/include/asm/pgtable.h 	return pte;
pte               342 arch/mips/include/asm/pgtable.h static inline pte_t pte_mkspecial(pte_t pte)
pte               344 arch/mips/include/asm/pgtable.h 	pte.pte_low |= _PAGE_SPECIAL;
pte               345 arch/mips/include/asm/pgtable.h 	return pte;
pte               348 arch/mips/include/asm/pgtable.h static inline int pte_write(pte_t pte)	{ return pte_val(pte) & _PAGE_WRITE; }
pte               349 arch/mips/include/asm/pgtable.h static inline int pte_dirty(pte_t pte)	{ return pte_val(pte) & _PAGE_MODIFIED; }
pte               350 arch/mips/include/asm/pgtable.h static inline int pte_young(pte_t pte)	{ return pte_val(pte) & _PAGE_ACCESSED; }
pte               351 arch/mips/include/asm/pgtable.h static inline int pte_special(pte_t pte) { return pte_val(pte) & _PAGE_SPECIAL; }
pte               353 arch/mips/include/asm/pgtable.h static inline pte_t pte_wrprotect(pte_t pte)
pte               355 arch/mips/include/asm/pgtable.h 	pte_val(pte) &= ~(_PAGE_WRITE | _PAGE_SILENT_WRITE);
pte               356 arch/mips/include/asm/pgtable.h 	return pte;
pte               359 arch/mips/include/asm/pgtable.h static inline pte_t pte_mkclean(pte_t pte)
pte               361 arch/mips/include/asm/pgtable.h 	pte_val(pte) &= ~(_PAGE_MODIFIED | _PAGE_SILENT_WRITE);
pte               362 arch/mips/include/asm/pgtable.h 	return pte;
pte               365 arch/mips/include/asm/pgtable.h static inline pte_t pte_mkold(pte_t pte)
pte               367 arch/mips/include/asm/pgtable.h 	pte_val(pte) &= ~(_PAGE_ACCESSED | _PAGE_SILENT_READ);
pte               368 arch/mips/include/asm/pgtable.h 	return pte;
pte               371 arch/mips/include/asm/pgtable.h static inline pte_t pte_mkwrite(pte_t pte)
pte               373 arch/mips/include/asm/pgtable.h 	pte_val(pte) |= _PAGE_WRITE;
pte               374 arch/mips/include/asm/pgtable.h 	if (pte_val(pte) & _PAGE_MODIFIED)
pte               375 arch/mips/include/asm/pgtable.h 		pte_val(pte) |= _PAGE_SILENT_WRITE;
pte               376 arch/mips/include/asm/pgtable.h 	return pte;
pte               379 arch/mips/include/asm/pgtable.h static inline pte_t pte_mkdirty(pte_t pte)
pte               381 arch/mips/include/asm/pgtable.h 	pte_val(pte) |= _PAGE_MODIFIED;
pte               382 arch/mips/include/asm/pgtable.h 	if (pte_val(pte) & _PAGE_WRITE)
pte               383 arch/mips/include/asm/pgtable.h 		pte_val(pte) |= _PAGE_SILENT_WRITE;
pte               384 arch/mips/include/asm/pgtable.h 	return pte;
pte               387 arch/mips/include/asm/pgtable.h static inline pte_t pte_mkyoung(pte_t pte)
pte               389 arch/mips/include/asm/pgtable.h 	pte_val(pte) |= _PAGE_ACCESSED;
pte               390 arch/mips/include/asm/pgtable.h 	if (!(pte_val(pte) & _PAGE_NO_READ))
pte               391 arch/mips/include/asm/pgtable.h 		pte_val(pte) |= _PAGE_SILENT_READ;
pte               392 arch/mips/include/asm/pgtable.h 	return pte;
pte               395 arch/mips/include/asm/pgtable.h static inline pte_t pte_mkspecial(pte_t pte)
pte               397 arch/mips/include/asm/pgtable.h 	pte_val(pte) |= _PAGE_SPECIAL;
pte               398 arch/mips/include/asm/pgtable.h 	return pte;
pte               402 arch/mips/include/asm/pgtable.h static inline int pte_huge(pte_t pte)	{ return pte_val(pte) & _PAGE_HUGE; }
pte               404 arch/mips/include/asm/pgtable.h static inline pte_t pte_mkhuge(pte_t pte)
pte               406 arch/mips/include/asm/pgtable.h 	pte_val(pte) |= _PAGE_HUGE;
pte               407 arch/mips/include/asm/pgtable.h 	return pte;
pte               448 arch/mips/include/asm/pgtable.h static inline pte_t pte_modify(pte_t pte, pgprot_t newprot)
pte               450 arch/mips/include/asm/pgtable.h 	pte.pte_low  &= (_PAGE_MODIFIED | _PAGE_ACCESSED | _PFNX_MASK);
pte               451 arch/mips/include/asm/pgtable.h 	pte.pte_high &= (_PFN_MASK | _CACHE_MASK);
pte               452 arch/mips/include/asm/pgtable.h 	pte.pte_low  |= pgprot_val(newprot) & ~_PFNX_MASK;
pte               453 arch/mips/include/asm/pgtable.h 	pte.pte_high |= pgprot_val(newprot) & ~(_PFN_MASK | _CACHE_MASK);
pte               454 arch/mips/include/asm/pgtable.h 	return pte;
pte               457 arch/mips/include/asm/pgtable.h static inline pte_t pte_modify(pte_t pte, pgprot_t newprot)
pte               459 arch/mips/include/asm/pgtable.h 	pte.pte_low  &= _PAGE_CHG_MASK;
pte               460 arch/mips/include/asm/pgtable.h 	pte.pte_high &= (_PFN_MASK | _CACHE_MASK);
pte               461 arch/mips/include/asm/pgtable.h 	pte.pte_low  |= pgprot_val(newprot);
pte               462 arch/mips/include/asm/pgtable.h 	pte.pte_high |= pgprot_val(newprot) & ~(_PFN_MASK | _CACHE_MASK);
pte               463 arch/mips/include/asm/pgtable.h 	return pte;
pte               466 arch/mips/include/asm/pgtable.h static inline pte_t pte_modify(pte_t pte, pgprot_t newprot)
pte               468 arch/mips/include/asm/pgtable.h 	return __pte((pte_val(pte) & _PAGE_CHG_MASK) |
pte               475 arch/mips/include/asm/pgtable.h 	pte_t pte);
pte               480 arch/mips/include/asm/pgtable.h 	pte_t pte = *ptep;
pte               481 arch/mips/include/asm/pgtable.h 	__update_tlb(vma, address, pte);
pte               487 arch/mips/include/asm/pgtable.h 	pte_t pte = *(pte_t *)pmdp;
pte               489 arch/mips/include/asm/pgtable.h 	__update_tlb(vma, address, pte);
pte               185 arch/mips/kvm/mmu.c static bool kvm_mips_flush_gpa_pte(pte_t *pte, unsigned long start_gpa,
pte               194 arch/mips/kvm/mmu.c 		if (!pte_present(pte[i]))
pte               197 arch/mips/kvm/mmu.c 		set_pte(pte + i, __pte(0));
pte               205 arch/mips/kvm/mmu.c 	pte_t *pte;
pte               216 arch/mips/kvm/mmu.c 		pte = pte_offset(pmd + i, 0);
pte               220 arch/mips/kvm/mmu.c 		if (kvm_mips_flush_gpa_pte(pte, start_gpa, end)) {
pte               222 arch/mips/kvm/mmu.c 			pte_free_kernel(NULL, pte);
pte               307 arch/mips/kvm/mmu.c static int kvm_mips_##name##_pte(pte_t *pte, unsigned long start,	\
pte               317 arch/mips/kvm/mmu.c 		if (!pte_present(pte[i]))				\
pte               320 arch/mips/kvm/mmu.c 		old = pte[i];						\
pte               324 arch/mips/kvm/mmu.c 		set_pte(pte + i, new);					\
pte               335 arch/mips/kvm/mmu.c 	pte_t *pte;							\
pte               345 arch/mips/kvm/mmu.c 		pte = pte_offset(pmd + i, 0);				\
pte               349 arch/mips/kvm/mmu.c 		ret |= kvm_mips_##name##_pte(pte, start, cur_end);	\
pte               554 arch/mips/kvm/mmu.c int kvm_set_spte_hva(struct kvm *kvm, unsigned long hva, pte_t pte)
pte               559 arch/mips/kvm/mmu.c 	ret = handle_hva_to_gpa(kvm, hva, end, &kvm_set_spte_handler, &pte);
pte               836 arch/mips/kvm/mmu.c static bool kvm_mips_flush_gva_pte(pte_t *pte, unsigned long start_gva,
pte               852 arch/mips/kvm/mmu.c 		if (!pte_present(pte[i]))
pte               855 arch/mips/kvm/mmu.c 		set_pte(pte + i, __pte(0));
pte               863 arch/mips/kvm/mmu.c 	pte_t *pte;
pte               874 arch/mips/kvm/mmu.c 		pte = pte_offset(pmd + i, 0);
pte               878 arch/mips/kvm/mmu.c 		if (kvm_mips_flush_gva_pte(pte, start_gva, end)) {
pte               880 arch/mips/kvm/mmu.c 			pte_free_kernel(NULL, pte);
pte               964 arch/mips/kvm/mmu.c static pte_t kvm_mips_gpa_pte_to_gva_unmapped(pte_t pte)
pte               971 arch/mips/kvm/mmu.c 	if (!pte_dirty(pte))
pte               972 arch/mips/kvm/mmu.c 		pte = pte_wrprotect(pte);
pte               974 arch/mips/kvm/mmu.c 	return pte;
pte               977 arch/mips/kvm/mmu.c static pte_t kvm_mips_gpa_pte_to_gva_mapped(pte_t pte, long entrylo)
pte               981 arch/mips/kvm/mmu.c 		pte = pte_mkclean(pte);
pte               983 arch/mips/kvm/mmu.c 	return kvm_mips_gpa_pte_to_gva_unmapped(pte);
pte               569 arch/mips/kvm/trap_emul.c 	pte_t *pte;
pte               595 arch/mips/kvm/trap_emul.c 				pte = pte_offset(pmd + k, 0);
pte               596 arch/mips/kvm/trap_emul.c 				pte_free_kernel(NULL, pte);
pte               130 arch/mips/mm/cache.c void __update_cache(unsigned long address, pte_t pte)
pte               134 arch/mips/mm/cache.c 	int exec = !pte_no_exec(pte) && !cpu_has_ic_fills_f_dc;
pte               136 arch/mips/mm/cache.c 	pfn = pte_pfn(pte);
pte                29 arch/mips/mm/hugetlbpage.c 	pte_t *pte = NULL;
pte                34 arch/mips/mm/hugetlbpage.c 		pte = (pte_t *)pmd_alloc(mm, pud, addr);
pte                36 arch/mips/mm/hugetlbpage.c 	return pte;
pte                90 arch/mips/mm/init.c 	pte_t pte;
pte               100 arch/mips/mm/init.c 	pte = mk_pte(page, prot);
pte               102 arch/mips/mm/init.c 	entrylo = pte_to_entrylo(pte.pte_high);
pte               104 arch/mips/mm/init.c 	entrylo = pte.pte_high;
pte               106 arch/mips/mm/init.c 	entrylo = pte_to_entrylo(pte_val(pte));
pte               120 arch/mips/mm/init.c 		entrylo = (pte.pte_low & _PFNX_MASK);
pte               237 arch/mips/mm/init.c 	pte_t *pte;
pte               253 arch/mips/mm/init.c 					pte = (pte_t *) memblock_alloc_low(PAGE_SIZE,
pte               255 arch/mips/mm/init.c 					if (!pte)
pte               260 arch/mips/mm/init.c 					set_pmd(pmd, __pmd((unsigned long)pte));
pte               261 arch/mips/mm/init.c 					BUG_ON(pte != pte_offset_kernel(pmd, 0));
pte                21 arch/mips/mm/ioremap.c static inline void remap_area_pte(pte_t * pte, unsigned long address,
pte                36 arch/mips/mm/ioremap.c 		if (!pte_none(*pte)) {
pte                40 arch/mips/mm/ioremap.c 		set_pte(pte, pfn_pte(pfn, pgprot));
pte                43 arch/mips/mm/ioremap.c 		pte++;
pte                59 arch/mips/mm/ioremap.c 		pte_t * pte = pte_alloc_kernel(pmd, address);
pte                60 arch/mips/mm/ioremap.c 		if (!pte)
pte                62 arch/mips/mm/ioremap.c 		remap_area_pte(pte, address, end - address, address + phys_addr, flags);
pte                61 arch/mips/mm/pgtable-32.c 	pte_t *pte;
pte                87 arch/mips/mm/pgtable-32.c 	pte = pte_offset_kernel(pmd, vaddr);
pte                88 arch/mips/mm/pgtable-32.c 	pkmap_page_table = pte;
pte               182 arch/mips/mm/tlb-r3k.c void __update_tlb(struct vm_area_struct *vma, unsigned long address, pte_t pte)
pte               209 arch/mips/mm/tlb-r3k.c 	write_c0_entrylo0(pte_val(pte));
pte               294 arch/mips/mm/tlb-r4k.c void __update_tlb(struct vm_area_struct * vma, unsigned long address, pte_t pte)
pte               731 arch/mips/mm/tlbex.c static void build_huge_update_entries(u32 **p, unsigned int pte,
pte               751 arch/mips/mm/tlbex.c 	build_convert_pte_to_entrylo(p, pte);
pte               752 arch/mips/mm/tlbex.c 	UASM_i_MTC0(p, pte, C0_ENTRYLO0); /* load it */
pte               755 arch/mips/mm/tlbex.c 		UASM_i_ADDIU(p, pte, pte, HPAGE_SIZE >> 7);
pte               757 arch/mips/mm/tlbex.c 		UASM_i_ADDU(p, pte, pte, tmp);
pte               759 arch/mips/mm/tlbex.c 	UASM_i_MTC0(p, pte, C0_ENTRYLO1); /* load it */
pte               764 arch/mips/mm/tlbex.c 				    unsigned int pte,
pte               769 arch/mips/mm/tlbex.c 	UASM_i_SC(p, pte, 0, ptr);
pte               770 arch/mips/mm/tlbex.c 	uasm_il_beqz(p, r, pte, label_tlb_huge_update);
pte               771 arch/mips/mm/tlbex.c 	UASM_i_LW(p, pte, 0, ptr); /* Needed because SC killed our PTE */
pte               773 arch/mips/mm/tlbex.c 	UASM_i_SW(p, pte, 0, ptr);
pte               785 arch/mips/mm/tlbex.c 		build_huge_update_entries(p, pte, ptr);
pte               786 arch/mips/mm/tlbex.c 		build_huge_tlb_write_entry(p, l, r, pte, tlb_random, 0);
pte               791 arch/mips/mm/tlbex.c 	build_huge_update_entries(p, pte, ptr);
pte               792 arch/mips/mm/tlbex.c 	build_huge_tlb_write_entry(p, l, r, pte, tlb_indexed, 0);
pte              1667 arch/mips/mm/tlbex.c iPTE_LW(u32 **p, unsigned int pte, unsigned int ptr)
pte              1674 arch/mips/mm/tlbex.c 		uasm_i_lld(p, pte, 0, ptr);
pte              1677 arch/mips/mm/tlbex.c 		UASM_i_LL(p, pte, 0, ptr);
pte              1681 arch/mips/mm/tlbex.c 		uasm_i_ld(p, pte, 0, ptr);
pte              1684 arch/mips/mm/tlbex.c 		UASM_i_LW(p, pte, 0, ptr);
pte              1689 arch/mips/mm/tlbex.c iPTE_SW(u32 **p, struct uasm_reloc **r, unsigned int pte, unsigned int ptr,
pte              1697 arch/mips/mm/tlbex.c 		uasm_i_or(p, pte, pte, scratch);
pte              1700 arch/mips/mm/tlbex.c 		uasm_i_ori(p, pte, pte, mode);
pte              1706 arch/mips/mm/tlbex.c 		uasm_i_scd(p, pte, 0, ptr);
pte              1709 arch/mips/mm/tlbex.c 		UASM_i_SC(p, pte, 0, ptr);
pte              1712 arch/mips/mm/tlbex.c 		uasm_il_beqzl(p, r, pte, label_smp_pgtable_change);
pte              1714 arch/mips/mm/tlbex.c 		uasm_il_beqz(p, r, pte, label_smp_pgtable_change);
pte              1719 arch/mips/mm/tlbex.c 		uasm_i_ll(p, pte, sizeof(pte_t) / 2, ptr);
pte              1720 arch/mips/mm/tlbex.c 		uasm_i_ori(p, pte, pte, hwmode);
pte              1722 arch/mips/mm/tlbex.c 		uasm_i_sc(p, pte, sizeof(pte_t) / 2, ptr);
pte              1723 arch/mips/mm/tlbex.c 		uasm_il_beqz(p, r, pte, label_smp_pgtable_change);
pte              1725 arch/mips/mm/tlbex.c 		uasm_i_lw(p, pte, 0, ptr);
pte              1734 arch/mips/mm/tlbex.c 		uasm_i_sd(p, pte, 0, ptr);
pte              1737 arch/mips/mm/tlbex.c 		UASM_i_SW(p, pte, 0, ptr);
pte              1741 arch/mips/mm/tlbex.c 		uasm_i_lw(p, pte, sizeof(pte_t) / 2, ptr);
pte              1742 arch/mips/mm/tlbex.c 		uasm_i_ori(p, pte, pte, hwmode);
pte              1744 arch/mips/mm/tlbex.c 		uasm_i_sw(p, pte, sizeof(pte_t) / 2, ptr);
pte              1745 arch/mips/mm/tlbex.c 		uasm_i_lw(p, pte, 0, ptr);
pte              1758 arch/mips/mm/tlbex.c 		  int pte, int ptr, int scratch, enum label_id lid)
pte              1760 arch/mips/mm/tlbex.c 	int t = scratch >= 0 ? scratch : pte;
pte              1761 arch/mips/mm/tlbex.c 	int cur = pte;
pte              1765 arch/mips/mm/tlbex.c 			uasm_il_bbit0(p, r, pte, ilog2(_PAGE_PRESENT), lid);
pte              1774 arch/mips/mm/tlbex.c 			if (pte == t)
pte              1776 arch/mips/mm/tlbex.c 				iPTE_LW(p, pte, ptr);
pte              1787 arch/mips/mm/tlbex.c 		if (pte == t)
pte              1789 arch/mips/mm/tlbex.c 			iPTE_LW(p, pte, ptr);
pte              1795 arch/mips/mm/tlbex.c build_make_valid(u32 **p, struct uasm_reloc **r, unsigned int pte,
pte              1800 arch/mips/mm/tlbex.c 	iPTE_SW(p, r, pte, ptr, mode, scratch);
pte              1809 arch/mips/mm/tlbex.c 		   unsigned int pte, unsigned int ptr, int scratch,
pte              1812 arch/mips/mm/tlbex.c 	int t = scratch >= 0 ? scratch : pte;
pte              1813 arch/mips/mm/tlbex.c 	int cur = pte;
pte              1824 arch/mips/mm/tlbex.c 	if (pte == t)
pte              1826 arch/mips/mm/tlbex.c 		iPTE_LW(p, pte, ptr);
pte              1835 arch/mips/mm/tlbex.c build_make_write(u32 **p, struct uasm_reloc **r, unsigned int pte,
pte              1841 arch/mips/mm/tlbex.c 	iPTE_SW(p, r, pte, ptr, mode, scratch);
pte              1850 arch/mips/mm/tlbex.c 		     unsigned int pte, unsigned int ptr, int scratch,
pte              1854 arch/mips/mm/tlbex.c 		uasm_il_bbit0(p, r, pte, ilog2(_PAGE_WRITE), lid);
pte              1857 arch/mips/mm/tlbex.c 		int t = scratch >= 0 ? scratch : pte;
pte              1858 arch/mips/mm/tlbex.c 		uasm_i_srl(p, t, pte, _PAGE_WRITE_SHIFT);
pte              1861 arch/mips/mm/tlbex.c 		if (pte == t)
pte              1863 arch/mips/mm/tlbex.c 			iPTE_LW(p, pte, ptr);
pte              1879 arch/mips/mm/tlbex.c build_r3000_pte_reload_tlbwi(u32 **p, unsigned int pte, unsigned int tmp)
pte              1881 arch/mips/mm/tlbex.c 	uasm_i_mtc0(p, pte, C0_ENTRYLO0); /* cp0 delay */
pte              1896 arch/mips/mm/tlbex.c 			     struct uasm_reloc **r, unsigned int pte,
pte              1900 arch/mips/mm/tlbex.c 	uasm_i_mtc0(p, pte, C0_ENTRYLO0); /* cp0 delay */
pte              1913 arch/mips/mm/tlbex.c build_r3000_tlbchange_handler_head(u32 **p, unsigned int pte,
pte              1918 arch/mips/mm/tlbex.c 	uasm_i_mfc0(p, pte, C0_BADVADDR);
pte              1921 arch/mips/mm/tlbex.c 	uasm_i_srl(p, pte, pte, 22); /* load delay */
pte              1922 arch/mips/mm/tlbex.c 	uasm_i_sll(p, pte, pte, 2);
pte              1923 arch/mips/mm/tlbex.c 	uasm_i_addu(p, ptr, ptr, pte);
pte              1924 arch/mips/mm/tlbex.c 	uasm_i_mfc0(p, pte, C0_CONTEXT);
pte              1926 arch/mips/mm/tlbex.c 	uasm_i_andi(p, pte, pte, 0xffc); /* load delay */
pte              1927 arch/mips/mm/tlbex.c 	uasm_i_addu(p, ptr, ptr, pte);
pte              1928 arch/mips/mm/tlbex.c 	uasm_i_lw(p, pte, 0, ptr);
pte               316 arch/mips/sgi-ip22/ip28-berr.c 			unsigned long pte = (lo >> 6) << 12; /* PTEBase */
pte               317 arch/mips/sgi-ip22/ip28-berr.c 			pte += 8*((vaddr >> pgsz) & 0x1ff);
pte               318 arch/mips/sgi-ip22/ip28-berr.c 			if (page_is_ram(PFN_DOWN(pte))) {
pte               325 arch/mips/sgi-ip22/ip28-berr.c 						PHYS_TO_XKSEG_UNCACHED(pte);
pte                20 arch/nds32/include/asm/pgalloc.h #define pgd_populate(mm, pmd, pte)	BUG()
pte                28 arch/nds32/include/asm/pgalloc.h 	pgtable_t pte;
pte                30 arch/nds32/include/asm/pgalloc.h 	pte = __pte_alloc_one(mm, GFP_PGTABLE_USER);
pte                31 arch/nds32/include/asm/pgalloc.h 	if (pte)
pte                32 arch/nds32/include/asm/pgalloc.h 		cpu_dcache_wb_page((unsigned long)page_address(pte));
pte                34 arch/nds32/include/asm/pgalloc.h 	return pte;
pte                40 arch/nds32/include/asm/pgtable.h #define pte_ERROR(pte)		__pte_error(__FILE__, __LINE__, pte_val(pte))
pte               187 arch/nds32/include/asm/pgtable.h #define pte_pfn(pte)		(pte_val(pte) >> PAGE_SHIFT)
pte               190 arch/nds32/include/asm/pgtable.h #define pte_none(pte)	        !(pte_val(pte))
pte               192 arch/nds32/include/asm/pgtable.h #define pte_page(pte)		(pfn_to_page(pte_pfn(pte)))
pte               200 arch/nds32/include/asm/pgtable.h #define pte_unmap(pte)		do { } while (0)
pte               201 arch/nds32/include/asm/pgtable.h #define pte_unmap_nested(pte)	do { } while (0)
pte               224 arch/nds32/include/asm/pgtable.h static inline void set_pte(pte_t * ptep, pte_t pte)
pte               227 arch/nds32/include/asm/pgtable.h 	*ptep = pte;
pte               249 arch/nds32/include/asm/pgtable.h #define pte_present(pte)        (pte_val(pte) & _PAGE_V)
pte               250 arch/nds32/include/asm/pgtable.h #define pte_write(pte)          ((pte_val(pte) & _PAGE_M_MASK) == _PAGE_M_URW_KRW)
pte               251 arch/nds32/include/asm/pgtable.h #define pte_read(pte)		(((pte_val(pte) & _PAGE_M_MASK) == _PAGE_M_UR_KR) || \
pte               252 arch/nds32/include/asm/pgtable.h 				((pte_val(pte) & _PAGE_M_MASK) == _PAGE_M_UR_KRW) || \
pte               253 arch/nds32/include/asm/pgtable.h 				((pte_val(pte) & _PAGE_M_MASK) == _PAGE_M_URW_KRW))
pte               254 arch/nds32/include/asm/pgtable.h #define pte_kernel_write(pte)   (((pte_val(pte) & _PAGE_M_MASK) == _PAGE_M_URW_KRW) || \
pte               255 arch/nds32/include/asm/pgtable.h 				((pte_val(pte) & _PAGE_M_MASK) == _PAGE_M_UR_KRW) || \
pte               256 arch/nds32/include/asm/pgtable.h 				((pte_val(pte) & _PAGE_M_MASK) == _PAGE_M_KRW) || \
pte               257 arch/nds32/include/asm/pgtable.h 				(((pte_val(pte) & _PAGE_M_MASK) == _PAGE_M_XKRW) && pte_exec(pte)))
pte               258 arch/nds32/include/asm/pgtable.h #define pte_exec(pte)		(pte_val(pte) & _PAGE_E)
pte               259 arch/nds32/include/asm/pgtable.h #define pte_dirty(pte)		(pte_val(pte) & _PAGE_D)
pte               260 arch/nds32/include/asm/pgtable.h #define pte_young(pte)		(pte_val(pte) & _PAGE_YOUNG)
pte               265 arch/nds32/include/asm/pgtable.h #define pte_file(pte)		(pte_val(pte) & _PAGE_FILE)
pte               272 arch/nds32/include/asm/pgtable.h static inline pte_t pte_##fn(pte_t pte) { pte_val(pte) op; return pte; }
pte               274 arch/nds32/include/asm/pgtable.h static inline pte_t pte_wrprotect(pte_t pte)
pte               276 arch/nds32/include/asm/pgtable.h 	pte_val(pte) = pte_val(pte) & ~_PAGE_M_MASK;
pte               277 arch/nds32/include/asm/pgtable.h 	pte_val(pte) = pte_val(pte) | _PAGE_M_UR_KR;
pte               278 arch/nds32/include/asm/pgtable.h 	return pte;
pte               281 arch/nds32/include/asm/pgtable.h static inline pte_t pte_mkwrite(pte_t pte)
pte               283 arch/nds32/include/asm/pgtable.h 	pte_val(pte) = pte_val(pte) & ~_PAGE_M_MASK;
pte               284 arch/nds32/include/asm/pgtable.h 	pte_val(pte) = pte_val(pte) | _PAGE_M_URW_KRW;
pte               285 arch/nds32/include/asm/pgtable.h 	return pte;
pte               294 arch/nds32/include/asm/pgtable.h static inline int pte_special(pte_t pte)
pte               299 arch/nds32/include/asm/pgtable.h static inline pte_t pte_mkspecial(pte_t pte)
pte               301 arch/nds32/include/asm/pgtable.h 	return pte;
pte               372 arch/nds32/include/asm/pgtable.h static inline pte_t pte_modify(pte_t pte, pgprot_t newprot)
pte               375 arch/nds32/include/asm/pgtable.h 	pte_val(pte) = (pte_val(pte) & ~mask) | (pgprot_val(newprot) & mask);
pte               376 arch/nds32/include/asm/pgtable.h 	return pte;
pte               388 arch/nds32/include/asm/pgtable.h #define __pte_to_swp_entry(pte)	     ((swp_entry_t) { pte_val(pte) })
pte                 9 arch/nds32/include/asm/tlb.h #define __pte_free_tlb(tlb, pte, addr)	pte_free((tlb)->mm, pte)
pte                44 arch/nds32/include/asm/tlbflush.h 		      unsigned long address, pte_t * pte);
pte                50 arch/nds32/mm/cacheflush.c 		      pte_t * pte)
pte                53 arch/nds32/mm/cacheflush.c 	unsigned long pfn = pte_pfn(*pte);
pte                62 arch/nds32/mm/cacheflush.c 		__nds32__tlbop_rwr(*pte);
pte                88 arch/nds32/mm/cacheflush.c 	unsigned long kaddr, pte;
pte                92 arch/nds32/mm/cacheflush.c 	pte = (pa | PAGE_KERNEL);
pte                94 arch/nds32/mm/cacheflush.c 	__nds32__tlbop_rwlk(pte);
pte               108 arch/nds32/mm/cacheflush.c 	unsigned long kaddr, pte;
pte               112 arch/nds32/mm/cacheflush.c 	pte = (pa | PAGE_KERNEL);
pte               114 arch/nds32/mm/cacheflush.c 	__nds32__tlbop_rwlk(pte);
pte                59 arch/nds32/mm/fault.c 			pte_t *pte;
pte                61 arch/nds32/mm/fault.c 			pte = pte_offset_map(pmd, addr);
pte                62 arch/nds32/mm/fault.c 			pr_alert(", *pte=%08lx", pte_val(*pte));
pte                63 arch/nds32/mm/fault.c 			pte_unmap(pte);
pte                38 arch/nds32/mm/highmem.c 	unsigned long vaddr, pte;
pte                51 arch/nds32/mm/highmem.c 	pte = (page_to_pfn(page) << PAGE_SHIFT) | (PAGE_KERNEL);
pte                53 arch/nds32/mm/highmem.c 	set_pte(ptep, pte);
pte                57 arch/nds32/mm/highmem.c 	__nds32__tlbop_rwr(pte);
pte                59 arch/nds32/mm/init.c 	pte_t *pte;
pte                81 arch/nds32/mm/init.c 		pte = memblock_alloc(PAGE_SIZE, PAGE_SIZE);
pte                82 arch/nds32/mm/init.c 		if (!pte)
pte                85 arch/nds32/mm/init.c 		set_pmd(pme, __pmd(__pa(pte) + _PAGE_KERNEL_TABLE));
pte                89 arch/nds32/mm/init.c 		     v += PAGE_SIZE, p += PAGE_SIZE, j++, pte++) {
pte                92 arch/nds32/mm/init.c 			set_pte(pte, __pte(p + pgprot_val(PAGE_KERNEL)));
pte               106 arch/nds32/mm/init.c 	pte_t *pte;
pte               131 arch/nds32/mm/init.c 	pte = memblock_alloc(PAGE_SIZE, PAGE_SIZE);
pte               132 arch/nds32/mm/init.c 	if (!pte)
pte               135 arch/nds32/mm/init.c 	set_pmd(pmd, __pmd(__pa(pte) + _PAGE_KERNEL_TABLE));
pte               136 arch/nds32/mm/init.c 	pkmap_page_table = pte;
pte               259 arch/nds32/mm/init.c 	pte_t *pte;
pte               263 arch/nds32/mm/init.c 	pte = (pte_t *)&fixmap_pmd_p[pte_index(addr)];
pte               266 arch/nds32/mm/init.c 		set_pte(pte, pfn_pte(phys >> PAGE_SHIFT, flags));
pte               268 arch/nds32/mm/init.c 		pte_clear(&init_mm, addr, pte);
pte                44 arch/nds32/mm/mm-nds32.c 	struct page *pte;
pte                58 arch/nds32/mm/mm-nds32.c 	pte = pmd_page(*pmd);
pte                61 arch/nds32/mm/mm-nds32.c 	pte_free(mm, pte);
pte                20 arch/nds32/mm/proc.c 	pte_t *ptep, pte;
pte                25 arch/nds32/mm/proc.c 		pte = *ptep;
pte                26 arch/nds32/mm/proc.c 		if (pte_present(pte))
pte                27 arch/nds32/mm/proc.c 			return pte;
pte                37 arch/nds32/mm/proc.c 	pte_t *ptep, pte;
pte                46 arch/nds32/mm/proc.c 				pte = *ptep;
pte                47 arch/nds32/mm/proc.c 				if (pte_present(pte))
pte                48 arch/nds32/mm/proc.c 					return pte;
pte                59 arch/nds32/mm/proc.c 	pte_t pte;
pte                64 arch/nds32/mm/proc.c 		pte = va_present(mm, addr);
pte                65 arch/nds32/mm/proc.c 		if (!pte && pte_read(pte))
pte                78 arch/nds32/mm/proc.c 	pte_t pte;
pte                83 arch/nds32/mm/proc.c 		pte = va_present(mm, addr);
pte                84 arch/nds32/mm/proc.c 		if (!pte && pte_write(pte))
pte                88 arch/nds32/mm/proc.c 		pte = va_kernel_present(addr);
pte                89 arch/nds32/mm/proc.c 		if (!pte && pte_kernel_write(pte))
pte                61 arch/nios2/include/asm/page.h typedef struct { unsigned long pte; } pte_t;
pte                65 arch/nios2/include/asm/page.h #define pte_val(x)	((x).pte)
pte                18 arch/nios2/include/asm/pgalloc.h 	pte_t *pte)
pte                20 arch/nios2/include/asm/pgalloc.h 	set_pmd(pmd, __pmd((unsigned long)pte));
pte                24 arch/nios2/include/asm/pgalloc.h 	pgtable_t pte)
pte                26 arch/nios2/include/asm/pgalloc.h 	set_pmd(pmd, __pmd((unsigned long)page_address(pte)));
pte                42 arch/nios2/include/asm/pgalloc.h #define __pte_free_tlb(tlb, pte, addr)				\
pte                44 arch/nios2/include/asm/pgalloc.h 		pgtable_pte_page_dtor(pte);			\
pte                45 arch/nios2/include/asm/pgalloc.h 		tlb_remove_page((tlb), (pte));			\
pte               110 arch/nios2/include/asm/pgtable.h static inline int pte_write(pte_t pte)		\
pte               111 arch/nios2/include/asm/pgtable.h 	{ return pte_val(pte) & _PAGE_WRITE; }
pte               112 arch/nios2/include/asm/pgtable.h static inline int pte_dirty(pte_t pte)		\
pte               113 arch/nios2/include/asm/pgtable.h 	{ return pte_val(pte) & _PAGE_DIRTY; }
pte               114 arch/nios2/include/asm/pgtable.h static inline int pte_young(pte_t pte)		\
pte               115 arch/nios2/include/asm/pgtable.h 	{ return pte_val(pte) & _PAGE_ACCESSED; }
pte               116 arch/nios2/include/asm/pgtable.h static inline int pte_special(pte_t pte)	{ return 0; }
pte               129 arch/nios2/include/asm/pgtable.h static inline int pte_none(pte_t pte)
pte               131 arch/nios2/include/asm/pgtable.h 	return !(pte_val(pte) & ~(_PAGE_GLOBAL|0xf));
pte               134 arch/nios2/include/asm/pgtable.h static inline int pte_present(pte_t pte)	\
pte               135 arch/nios2/include/asm/pgtable.h 	{ return pte_val(pte) & _PAGE_PRESENT; }
pte               141 arch/nios2/include/asm/pgtable.h static inline pte_t pte_wrprotect(pte_t pte)
pte               143 arch/nios2/include/asm/pgtable.h 	pte_val(pte) &= ~_PAGE_WRITE;
pte               144 arch/nios2/include/asm/pgtable.h 	return pte;
pte               147 arch/nios2/include/asm/pgtable.h static inline pte_t pte_mkclean(pte_t pte)
pte               149 arch/nios2/include/asm/pgtable.h 	pte_val(pte) &= ~_PAGE_DIRTY;
pte               150 arch/nios2/include/asm/pgtable.h 	return pte;
pte               153 arch/nios2/include/asm/pgtable.h static inline pte_t pte_mkold(pte_t pte)
pte               155 arch/nios2/include/asm/pgtable.h 	pte_val(pte) &= ~_PAGE_ACCESSED;
pte               156 arch/nios2/include/asm/pgtable.h 	return pte;
pte               159 arch/nios2/include/asm/pgtable.h static inline pte_t pte_mkwrite(pte_t pte)
pte               161 arch/nios2/include/asm/pgtable.h 	pte_val(pte) |= _PAGE_WRITE;
pte               162 arch/nios2/include/asm/pgtable.h 	return pte;
pte               165 arch/nios2/include/asm/pgtable.h static inline pte_t pte_mkdirty(pte_t pte)
pte               167 arch/nios2/include/asm/pgtable.h 	pte_val(pte) |= _PAGE_DIRTY;
pte               168 arch/nios2/include/asm/pgtable.h 	return pte;
pte               171 arch/nios2/include/asm/pgtable.h static inline pte_t pte_mkspecial(pte_t pte)	{ return pte; }
pte               173 arch/nios2/include/asm/pgtable.h static inline pte_t pte_mkyoung(pte_t pte)
pte               175 arch/nios2/include/asm/pgtable.h 	pte_val(pte) |= _PAGE_ACCESSED;
pte               176 arch/nios2/include/asm/pgtable.h 	return pte;
pte               179 arch/nios2/include/asm/pgtable.h static inline pte_t pte_modify(pte_t pte, pgprot_t newprot)
pte               183 arch/nios2/include/asm/pgtable.h 	pte_val(pte) = (pte_val(pte) & ~mask) | (pgprot_val(newprot) & mask);
pte               184 arch/nios2/include/asm/pgtable.h 	return pte;
pte               198 arch/nios2/include/asm/pgtable.h #define pte_pfn(pte)		(pte_val(pte) & 0xfffff)
pte               200 arch/nios2/include/asm/pgtable.h #define pte_page(pte)		(pfn_to_page(pte_pfn(pte)))
pte               243 arch/nios2/include/asm/pgtable.h #define pte_unmap(pte)	do { } while (0)
pte               288 arch/nios2/include/asm/pgtable.h #define __pte_to_swp_entry(pte)	((swp_entry_t) { pte_val(pte) })
pte               298 arch/nios2/include/asm/pgtable.h 			     unsigned long address, pte_t *pte);
pte                42 arch/nios2/include/asm/tlbflush.h 			    pte_t pte);
pte               203 arch/nios2/mm/cacheflush.c 	pte_t pte = *ptep;
pte               204 arch/nios2/mm/cacheflush.c 	unsigned long pfn = pte_pfn(pte);
pte               208 arch/nios2/mm/cacheflush.c 	reload_tlb_page(vma, address, pte);
pte                22 arch/nios2/mm/ioremap.c static inline void remap_area_pte(pte_t *pte, unsigned long address,
pte                39 arch/nios2/mm/ioremap.c 		if (!pte_none(*pte)) {
pte                43 arch/nios2/mm/ioremap.c 		set_pte(pte, pfn_pte(pfn, pgprot));
pte                46 arch/nios2/mm/ioremap.c 		pte++;
pte                64 arch/nios2/mm/ioremap.c 		pte_t *pte = pte_alloc_kernel(pmd, address);
pte                66 arch/nios2/mm/ioremap.c 		if (!pte)
pte                68 arch/nios2/mm/ioremap.c 		remap_area_pte(pte, address, end - address, address + phys_addr,
pte                96 arch/nios2/mm/tlb.c static void reload_tlb_one_pid(unsigned long addr, unsigned long mmu_pid, pte_t pte)
pte               100 arch/nios2/mm/tlb.c 	replace_tlb_one_pid(addr, mmu_pid, pte_val(pte));
pte               114 arch/nios2/mm/tlb.c void reload_tlb_page(struct vm_area_struct *vma, unsigned long addr, pte_t pte)
pte               118 arch/nios2/mm/tlb.c 	reload_tlb_one_pid(addr, mmu_pid, pte);
pte                49 arch/openrisc/include/asm/page.h 	unsigned long pte;
pte                59 arch/openrisc/include/asm/page.h #define pte_val(x)	((x).pte)
pte                25 arch/openrisc/include/asm/pgalloc.h #define pmd_populate_kernel(mm, pmd, pte) \
pte                26 arch/openrisc/include/asm/pgalloc.h 	set_pmd(pmd, __pmd(_KERNPG_TABLE + __pa(pte)))
pte                29 arch/openrisc/include/asm/pgalloc.h 				struct page *pte)
pte                32 arch/openrisc/include/asm/pgalloc.h 		     ((unsigned long)page_to_pfn(pte) <<
pte                73 arch/openrisc/include/asm/pgalloc.h 	struct page *pte;
pte                74 arch/openrisc/include/asm/pgalloc.h 	pte = alloc_pages(GFP_KERNEL, 0);
pte                75 arch/openrisc/include/asm/pgalloc.h 	if (!pte)
pte                77 arch/openrisc/include/asm/pgalloc.h 	clear_page(page_address(pte));
pte                78 arch/openrisc/include/asm/pgalloc.h 	if (!pgtable_pte_page_ctor(pte)) {
pte                79 arch/openrisc/include/asm/pgalloc.h 		__free_page(pte);
pte                82 arch/openrisc/include/asm/pgalloc.h 	return pte;
pte                85 arch/openrisc/include/asm/pgalloc.h static inline void pte_free_kernel(struct mm_struct *mm, pte_t *pte)
pte                87 arch/openrisc/include/asm/pgalloc.h 	free_page((unsigned long)pte);
pte                90 arch/openrisc/include/asm/pgalloc.h static inline void pte_free(struct mm_struct *mm, struct page *pte)
pte                92 arch/openrisc/include/asm/pgalloc.h 	pgtable_pte_page_dtor(pte);
pte                93 arch/openrisc/include/asm/pgalloc.h 	__free_page(pte);
pte                96 arch/openrisc/include/asm/pgalloc.h #define __pte_free_tlb(tlb, pte, addr)	\
pte                98 arch/openrisc/include/asm/pgalloc.h 	pgtable_pte_page_dtor(pte);	\
pte                99 arch/openrisc/include/asm/pgalloc.h 	tlb_remove_page((tlb), (pte));	\
pte               234 arch/openrisc/include/asm/pgtable.h static inline int pte_read(pte_t pte)  { return pte_val(pte) & _PAGE_READ; }
pte               235 arch/openrisc/include/asm/pgtable.h static inline int pte_write(pte_t pte) { return pte_val(pte) & _PAGE_WRITE; }
pte               236 arch/openrisc/include/asm/pgtable.h static inline int pte_exec(pte_t pte)  { return pte_val(pte) & _PAGE_EXEC; }
pte               237 arch/openrisc/include/asm/pgtable.h static inline int pte_dirty(pte_t pte) { return pte_val(pte) & _PAGE_DIRTY; }
pte               238 arch/openrisc/include/asm/pgtable.h static inline int pte_young(pte_t pte) { return pte_val(pte) & _PAGE_ACCESSED; }
pte               239 arch/openrisc/include/asm/pgtable.h static inline int pte_special(pte_t pte) { return 0; }
pte               240 arch/openrisc/include/asm/pgtable.h static inline pte_t pte_mkspecial(pte_t pte) { return pte; }
pte               242 arch/openrisc/include/asm/pgtable.h static inline pte_t pte_wrprotect(pte_t pte)
pte               244 arch/openrisc/include/asm/pgtable.h 	pte_val(pte) &= ~(_PAGE_WRITE);
pte               245 arch/openrisc/include/asm/pgtable.h 	return pte;
pte               248 arch/openrisc/include/asm/pgtable.h static inline pte_t pte_rdprotect(pte_t pte)
pte               250 arch/openrisc/include/asm/pgtable.h 	pte_val(pte) &= ~(_PAGE_READ);
pte               251 arch/openrisc/include/asm/pgtable.h 	return pte;
pte               254 arch/openrisc/include/asm/pgtable.h static inline pte_t pte_exprotect(pte_t pte)
pte               256 arch/openrisc/include/asm/pgtable.h 	pte_val(pte) &= ~(_PAGE_EXEC);
pte               257 arch/openrisc/include/asm/pgtable.h 	return pte;
pte               260 arch/openrisc/include/asm/pgtable.h static inline pte_t pte_mkclean(pte_t pte)
pte               262 arch/openrisc/include/asm/pgtable.h 	pte_val(pte) &= ~(_PAGE_DIRTY);
pte               263 arch/openrisc/include/asm/pgtable.h 	return pte;
pte               266 arch/openrisc/include/asm/pgtable.h static inline pte_t pte_mkold(pte_t pte)
pte               268 arch/openrisc/include/asm/pgtable.h 	pte_val(pte) &= ~(_PAGE_ACCESSED);
pte               269 arch/openrisc/include/asm/pgtable.h 	return pte;
pte               272 arch/openrisc/include/asm/pgtable.h static inline pte_t pte_mkwrite(pte_t pte)
pte               274 arch/openrisc/include/asm/pgtable.h 	pte_val(pte) |= _PAGE_WRITE;
pte               275 arch/openrisc/include/asm/pgtable.h 	return pte;
pte               278 arch/openrisc/include/asm/pgtable.h static inline pte_t pte_mkread(pte_t pte)
pte               280 arch/openrisc/include/asm/pgtable.h 	pte_val(pte) |= _PAGE_READ;
pte               281 arch/openrisc/include/asm/pgtable.h 	return pte;
pte               284 arch/openrisc/include/asm/pgtable.h static inline pte_t pte_mkexec(pte_t pte)
pte               286 arch/openrisc/include/asm/pgtable.h 	pte_val(pte) |= _PAGE_EXEC;
pte               287 arch/openrisc/include/asm/pgtable.h 	return pte;
pte               290 arch/openrisc/include/asm/pgtable.h static inline pte_t pte_mkdirty(pte_t pte)
pte               292 arch/openrisc/include/asm/pgtable.h 	pte_val(pte) |= _PAGE_DIRTY;
pte               293 arch/openrisc/include/asm/pgtable.h 	return pte;
pte               296 arch/openrisc/include/asm/pgtable.h static inline pte_t pte_mkyoung(pte_t pte)
pte               298 arch/openrisc/include/asm/pgtable.h 	pte_val(pte) |= _PAGE_ACCESSED;
pte               299 arch/openrisc/include/asm/pgtable.h 	return pte;
pte               315 arch/openrisc/include/asm/pgtable.h 	pte_t pte;
pte               317 arch/openrisc/include/asm/pgtable.h 	pte_val(pte) = __pa(page) | pgprot_val(pgprot);
pte               318 arch/openrisc/include/asm/pgtable.h 	return pte;
pte               331 arch/openrisc/include/asm/pgtable.h static inline pte_t pte_modify(pte_t pte, pgprot_t newprot)
pte               333 arch/openrisc/include/asm/pgtable.h 	pte_val(pte) = (pte_val(pte) & _PAGE_CHG_MASK) | pgprot_val(newprot);
pte               334 arch/openrisc/include/asm/pgtable.h 	return pte;
pte               345 arch/openrisc/include/asm/pgtable.h static inline unsigned long __pte_page(pte_t pte)
pte               348 arch/openrisc/include/asm/pgtable.h 	return (unsigned long)__va(pte_val(pte) & PAGE_MASK);
pte               351 arch/openrisc/include/asm/pgtable.h #define pte_pagenr(pte)         ((__pte_page(pte) - PAGE_OFFSET) >> PAGE_SHIFT)
pte               356 arch/openrisc/include/asm/pgtable.h #define pte_page(pte)		(mem_map+pte_pagenr(pte))
pte               399 arch/openrisc/include/asm/pgtable.h #define pte_unmap(pte)          do { } while (0)
pte               400 arch/openrisc/include/asm/pgtable.h #define pte_unmap_nested(pte)   do { } while (0)
pte               401 arch/openrisc/include/asm/pgtable.h #define pte_pfn(x)		((unsigned long)(((x).pte)) >> PAGE_SHIFT)
pte               416 arch/openrisc/include/asm/pgtable.h 	unsigned long address, pte_t *pte)
pte               421 arch/openrisc/include/asm/pgtable.h 	unsigned long address, pte_t *pte);
pte               424 arch/openrisc/include/asm/pgtable.h 	unsigned long address, pte_t *pte)
pte               426 arch/openrisc/include/asm/pgtable.h 	update_tlb(vma, address, pte);
pte               427 arch/openrisc/include/asm/pgtable.h 	update_cache(vma, address, pte);
pte               439 arch/openrisc/include/asm/pgtable.h #define __pte_to_swp_entry(pte)		((swp_entry_t) { pte_val(pte) })
pte                26 arch/openrisc/kernel/dma.c page_set_nocache(pte_t *pte, unsigned long addr,
pte                32 arch/openrisc/kernel/dma.c 	pte_val(*pte) |= _PAGE_CI;
pte                52 arch/openrisc/kernel/dma.c page_clear_nocache(pte_t *pte, unsigned long addr,
pte                55 arch/openrisc/kernel/dma.c 	pte_val(*pte) &= ~_PAGE_CI;
pte                43 arch/openrisc/mm/cache.c 	pte_t *pte)
pte                45 arch/openrisc/mm/cache.c 	unsigned long pfn = pte_val(*pte) >> PAGE_SHIFT;
pte                76 arch/openrisc/mm/init.c 	pte_t *pte;
pte               103 arch/openrisc/mm/init.c 			pte = memblock_alloc_raw(PAGE_SIZE, PAGE_SIZE);
pte               104 arch/openrisc/mm/init.c 			if (!pte)
pte               107 arch/openrisc/mm/init.c 			set_pmd(pme, __pmd(_KERNPG_TABLE + __pa(pte)));
pte               111 arch/openrisc/mm/init.c 			     v += PAGE_SIZE, p += PAGE_SIZE, j++, pte++) {
pte               118 arch/openrisc/mm/init.c 				set_pte(pte, mk_pte_phys(p, prot));
pte               119 arch/openrisc/mm/ioremap.c 	pte_t *pte;
pte               122 arch/openrisc/mm/ioremap.c 		pte = (pte_t *)get_zeroed_page(GFP_KERNEL);
pte               124 arch/openrisc/mm/ioremap.c 		pte = memblock_alloc(PAGE_SIZE, PAGE_SIZE);
pte               125 arch/openrisc/mm/ioremap.c 		if (!pte)
pte               130 arch/openrisc/mm/ioremap.c 	return pte;
pte                 9 arch/parisc/include/asm/hugetlb.h 		     pte_t *ptep, pte_t pte);
pte                49 arch/parisc/include/asm/hugetlb.h 					     pte_t pte, int dirty);
pte                41 arch/parisc/include/asm/page.h typedef struct { unsigned long pte; } pte_t; /* either 32 or 64bit */
pte                49 arch/parisc/include/asm/page.h #define pte_val(x)	((x).pte)
pte               102 arch/parisc/include/asm/pgalloc.h #define pgd_populate(mm, pmd, pte)	BUG()
pte               107 arch/parisc/include/asm/pgalloc.h pmd_populate_kernel(struct mm_struct *mm, pmd_t *pmd, pte_t *pte)
pte               116 arch/parisc/include/asm/pgalloc.h 			+ (__u32)(__pa((unsigned long)pte) >> PxD_VALUE_SHIFT));
pte               120 arch/parisc/include/asm/pgalloc.h 			+ (__u32)(__pa((unsigned long)pte) >> PxD_VALUE_SHIFT));
pte               379 arch/parisc/include/asm/pgtable.h static inline int pte_dirty(pte_t pte)		{ return pte_val(pte) & _PAGE_DIRTY; }
pte               380 arch/parisc/include/asm/pgtable.h static inline int pte_young(pte_t pte)		{ return pte_val(pte) & _PAGE_ACCESSED; }
pte               381 arch/parisc/include/asm/pgtable.h static inline int pte_write(pte_t pte)		{ return pte_val(pte) & _PAGE_WRITE; }
pte               382 arch/parisc/include/asm/pgtable.h static inline int pte_special(pte_t pte)	{ return 0; }
pte               384 arch/parisc/include/asm/pgtable.h static inline pte_t pte_mkclean(pte_t pte)	{ pte_val(pte) &= ~_PAGE_DIRTY; return pte; }
pte               385 arch/parisc/include/asm/pgtable.h static inline pte_t pte_mkold(pte_t pte)	{ pte_val(pte) &= ~_PAGE_ACCESSED; return pte; }
pte               386 arch/parisc/include/asm/pgtable.h static inline pte_t pte_wrprotect(pte_t pte)	{ pte_val(pte) &= ~_PAGE_WRITE; return pte; }
pte               387 arch/parisc/include/asm/pgtable.h static inline pte_t pte_mkdirty(pte_t pte)	{ pte_val(pte) |= _PAGE_DIRTY; return pte; }
pte               388 arch/parisc/include/asm/pgtable.h static inline pte_t pte_mkyoung(pte_t pte)	{ pte_val(pte) |= _PAGE_ACCESSED; return pte; }
pte               389 arch/parisc/include/asm/pgtable.h static inline pte_t pte_mkwrite(pte_t pte)	{ pte_val(pte) |= _PAGE_WRITE; return pte; }
pte               390 arch/parisc/include/asm/pgtable.h static inline pte_t pte_mkspecial(pte_t pte)	{ return pte; }
pte               396 arch/parisc/include/asm/pgtable.h #define pte_huge(pte)           (pte_val(pte) & _PAGE_HUGE)
pte               397 arch/parisc/include/asm/pgtable.h #define pte_mkhuge(pte)         (__pte(pte_val(pte) | \
pte               400 arch/parisc/include/asm/pgtable.h #define pte_huge(pte)           (0)
pte               401 arch/parisc/include/asm/pgtable.h #define pte_mkhuge(pte)         (pte)
pte               422 arch/parisc/include/asm/pgtable.h 	pte_t pte;
pte               423 arch/parisc/include/asm/pgtable.h 	pte_val(pte) = (pfn << PFN_PTE_SHIFT) | pgprot_val(pgprot);
pte               424 arch/parisc/include/asm/pgtable.h 	return pte;
pte               427 arch/parisc/include/asm/pgtable.h static inline pte_t pte_modify(pte_t pte, pgprot_t newprot)
pte               428 arch/parisc/include/asm/pgtable.h { pte_val(pte) = (pte_val(pte) & _PAGE_CHG_MASK) | pgprot_val(newprot); return pte; }
pte               434 arch/parisc/include/asm/pgtable.h #define pte_page(pte)		(pfn_to_page(pte_pfn(pte)))
pte               465 arch/parisc/include/asm/pgtable.h #define pte_unmap(pte) do { } while (0)
pte               467 arch/parisc/include/asm/pgtable.h #define pte_unmap(pte)			do { } while (0)
pte               468 arch/parisc/include/asm/pgtable.h #define pte_unmap_nested(pte)		do { } while (0)
pte               486 arch/parisc/include/asm/pgtable.h #define __pte_to_swp_entry(pte)		((swp_entry_t) { pte_val(pte) })
pte               500 arch/parisc/include/asm/pgtable.h 	pte_t pte;
pte               507 arch/parisc/include/asm/pgtable.h 	pte = *ptep;
pte               508 arch/parisc/include/asm/pgtable.h 	if (!pte_young(pte)) {
pte               512 arch/parisc/include/asm/pgtable.h 	set_pte(ptep, pte_mkold(pte));
pte                 8 arch/parisc/include/asm/tlb.h #define __pte_free_tlb(tlb, pte, addr)	pte_free((tlb)->mm, pte)
pte                77 arch/parisc/kernel/pci-dma.c static inline int map_pte_uncached(pte_t * pte,
pte                91 arch/parisc/kernel/pci-dma.c 		if (!pte_none(*pte))
pte                94 arch/parisc/kernel/pci-dma.c 		set_pte(pte, __mk_pte(*paddr_ptr, PAGE_KERNEL_UNC));
pte               100 arch/parisc/kernel/pci-dma.c 		pte++;
pte               116 arch/parisc/kernel/pci-dma.c 		pte_t * pte = pte_alloc_kernel(pmd, vaddr);
pte               117 arch/parisc/kernel/pci-dma.c 		if (!pte)
pte               119 arch/parisc/kernel/pci-dma.c 		if (map_pte_uncached(pte, orig_vaddr, end - vaddr, paddr_ptr))
pte               152 arch/parisc/kernel/pci-dma.c 	pte_t * pte;
pte               163 arch/parisc/kernel/pci-dma.c 	pte = pte_offset_map(pmd, vaddr);
pte               170 arch/parisc/kernel/pci-dma.c 		pte_t page = *pte;
pte               172 arch/parisc/kernel/pci-dma.c 		pte_clear(&init_mm, vaddr, pte);
pte               178 arch/parisc/kernel/pci-dma.c 		pte++;
pte                18 arch/parisc/mm/fixmap.c 	pte_t *pte;
pte                23 arch/parisc/mm/fixmap.c 	pte = pte_offset_kernel(pmd, vaddr);
pte                24 arch/parisc/mm/fixmap.c 	if (pte_none(*pte))
pte                25 arch/parisc/mm/fixmap.c 		pte = pte_alloc_kernel(pmd, vaddr);
pte                27 arch/parisc/mm/fixmap.c 	set_pte_at(&init_mm, vaddr, pte, __mk_pte(phys, PAGE_KERNEL_RWX));
pte                36 arch/parisc/mm/fixmap.c 	pte_t *pte = pte_offset_kernel(pmd, vaddr);
pte                38 arch/parisc/mm/fixmap.c 	if (WARN_ON(pte_none(*pte)))
pte                41 arch/parisc/mm/fixmap.c 	pte_clear(&init_mm, vaddr, pte);
pte                54 arch/parisc/mm/hugetlbpage.c 	pte_t *pte = NULL;
pte                68 arch/parisc/mm/hugetlbpage.c 			pte = pte_alloc_map(mm, pmd, addr);
pte                70 arch/parisc/mm/hugetlbpage.c 	return pte;
pte                79 arch/parisc/mm/hugetlbpage.c 	pte_t *pte = NULL;
pte                89 arch/parisc/mm/hugetlbpage.c 				pte = pte_offset_map(pmd, addr);
pte                92 arch/parisc/mm/hugetlbpage.c 	return pte;
pte               177 arch/parisc/mm/hugetlbpage.c 				pte_t pte, int dirty)
pte               184 arch/parisc/mm/hugetlbpage.c 	changed = !pte_same(*ptep, pte);
pte               186 arch/parisc/mm/hugetlbpage.c 		__set_huge_pte_at(mm, addr, ptep, pte);
pte               432 arch/parisc/mm/init.c 				pte_t pte;
pte               454 arch/parisc/mm/init.c 				pte = __mk_pte(address, prot);
pte               456 arch/parisc/mm/init.c 					pte = pte_mkhuge(pte);
pte               461 arch/parisc/mm/init.c 				set_pte(pg_table, pte);
pte                29 arch/powerpc/include/asm/book3s/32/pgalloc.h 				       pte_t *pte)
pte                31 arch/powerpc/include/asm/book3s/32/pgalloc.h 	*pmdp = __pmd(__pa(pte) | _PMD_PRESENT);
pte                21 arch/powerpc/include/asm/book3s/32/pgtable.h static inline bool pte_user(pte_t pte)
pte                23 arch/powerpc/include/asm/book3s/32/pgtable.h 	return pte_val(pte) & _PAGE_USER;
pte               367 arch/powerpc/include/asm/book3s/32/pgtable.h #define pte_unmap(pte)		kunmap_atomic(pte)
pte               378 arch/powerpc/include/asm/book3s/32/pgtable.h #define __pte_to_swp_entry(pte)		((swp_entry_t) { pte_val(pte) >> 3 })
pte               382 arch/powerpc/include/asm/book3s/32/pgtable.h static inline int pte_write(pte_t pte)		{ return !!(pte_val(pte) & _PAGE_RW);}
pte               383 arch/powerpc/include/asm/book3s/32/pgtable.h static inline int pte_read(pte_t pte)		{ return 1; }
pte               384 arch/powerpc/include/asm/book3s/32/pgtable.h static inline int pte_dirty(pte_t pte)		{ return !!(pte_val(pte) & _PAGE_DIRTY); }
pte               385 arch/powerpc/include/asm/book3s/32/pgtable.h static inline int pte_young(pte_t pte)		{ return !!(pte_val(pte) & _PAGE_ACCESSED); }
pte               386 arch/powerpc/include/asm/book3s/32/pgtable.h static inline int pte_special(pte_t pte)	{ return !!(pte_val(pte) & _PAGE_SPECIAL); }
pte               387 arch/powerpc/include/asm/book3s/32/pgtable.h static inline int pte_none(pte_t pte)		{ return (pte_val(pte) & ~_PTE_NONE_MASK) == 0; }
pte               388 arch/powerpc/include/asm/book3s/32/pgtable.h static inline bool pte_exec(pte_t pte)		{ return pte_val(pte) & _PAGE_EXEC; }
pte               390 arch/powerpc/include/asm/book3s/32/pgtable.h static inline int pte_present(pte_t pte)
pte               392 arch/powerpc/include/asm/book3s/32/pgtable.h 	return pte_val(pte) & _PAGE_PRESENT;
pte               395 arch/powerpc/include/asm/book3s/32/pgtable.h static inline bool pte_hw_valid(pte_t pte)
pte               397 arch/powerpc/include/asm/book3s/32/pgtable.h 	return pte_val(pte) & _PAGE_PRESENT;
pte               400 arch/powerpc/include/asm/book3s/32/pgtable.h static inline bool pte_hashpte(pte_t pte)
pte               402 arch/powerpc/include/asm/book3s/32/pgtable.h 	return !!(pte_val(pte) & _PAGE_HASHPTE);
pte               405 arch/powerpc/include/asm/book3s/32/pgtable.h static inline bool pte_ci(pte_t pte)
pte               407 arch/powerpc/include/asm/book3s/32/pgtable.h 	return !!(pte_val(pte) & _PAGE_NO_CACHE);
pte               415 arch/powerpc/include/asm/book3s/32/pgtable.h static inline bool pte_access_permitted(pte_t pte, bool write)
pte               421 arch/powerpc/include/asm/book3s/32/pgtable.h 	if (!pte_present(pte) || !pte_user(pte) || !pte_read(pte))
pte               424 arch/powerpc/include/asm/book3s/32/pgtable.h 	if (write && !pte_write(pte))
pte               442 arch/powerpc/include/asm/book3s/32/pgtable.h static inline unsigned long pte_pfn(pte_t pte)
pte               444 arch/powerpc/include/asm/book3s/32/pgtable.h 	return pte_val(pte) >> PTE_RPN_SHIFT;
pte               448 arch/powerpc/include/asm/book3s/32/pgtable.h static inline pte_t pte_wrprotect(pte_t pte)
pte               450 arch/powerpc/include/asm/book3s/32/pgtable.h 	return __pte(pte_val(pte) & ~_PAGE_RW);
pte               453 arch/powerpc/include/asm/book3s/32/pgtable.h static inline pte_t pte_exprotect(pte_t pte)
pte               455 arch/powerpc/include/asm/book3s/32/pgtable.h 	return __pte(pte_val(pte) & ~_PAGE_EXEC);
pte               458 arch/powerpc/include/asm/book3s/32/pgtable.h static inline pte_t pte_mkclean(pte_t pte)
pte               460 arch/powerpc/include/asm/book3s/32/pgtable.h 	return __pte(pte_val(pte) & ~_PAGE_DIRTY);
pte               463 arch/powerpc/include/asm/book3s/32/pgtable.h static inline pte_t pte_mkold(pte_t pte)
pte               465 arch/powerpc/include/asm/book3s/32/pgtable.h 	return __pte(pte_val(pte) & ~_PAGE_ACCESSED);
pte               468 arch/powerpc/include/asm/book3s/32/pgtable.h static inline pte_t pte_mkexec(pte_t pte)
pte               470 arch/powerpc/include/asm/book3s/32/pgtable.h 	return __pte(pte_val(pte) | _PAGE_EXEC);
pte               473 arch/powerpc/include/asm/book3s/32/pgtable.h static inline pte_t pte_mkpte(pte_t pte)
pte               475 arch/powerpc/include/asm/book3s/32/pgtable.h 	return pte;
pte               478 arch/powerpc/include/asm/book3s/32/pgtable.h static inline pte_t pte_mkwrite(pte_t pte)
pte               480 arch/powerpc/include/asm/book3s/32/pgtable.h 	return __pte(pte_val(pte) | _PAGE_RW);
pte               483 arch/powerpc/include/asm/book3s/32/pgtable.h static inline pte_t pte_mkdirty(pte_t pte)
pte               485 arch/powerpc/include/asm/book3s/32/pgtable.h 	return __pte(pte_val(pte) | _PAGE_DIRTY);
pte               488 arch/powerpc/include/asm/book3s/32/pgtable.h static inline pte_t pte_mkyoung(pte_t pte)
pte               490 arch/powerpc/include/asm/book3s/32/pgtable.h 	return __pte(pte_val(pte) | _PAGE_ACCESSED);
pte               493 arch/powerpc/include/asm/book3s/32/pgtable.h static inline pte_t pte_mkspecial(pte_t pte)
pte               495 arch/powerpc/include/asm/book3s/32/pgtable.h 	return __pte(pte_val(pte) | _PAGE_SPECIAL);
pte               498 arch/powerpc/include/asm/book3s/32/pgtable.h static inline pte_t pte_mkhuge(pte_t pte)
pte               500 arch/powerpc/include/asm/book3s/32/pgtable.h 	return pte;
pte               503 arch/powerpc/include/asm/book3s/32/pgtable.h static inline pte_t pte_mkprivileged(pte_t pte)
pte               505 arch/powerpc/include/asm/book3s/32/pgtable.h 	return __pte(pte_val(pte) & ~_PAGE_USER);
pte               508 arch/powerpc/include/asm/book3s/32/pgtable.h static inline pte_t pte_mkuser(pte_t pte)
pte               510 arch/powerpc/include/asm/book3s/32/pgtable.h 	return __pte(pte_val(pte) | _PAGE_USER);
pte               513 arch/powerpc/include/asm/book3s/32/pgtable.h static inline pte_t pte_modify(pte_t pte, pgprot_t newprot)
pte               515 arch/powerpc/include/asm/book3s/32/pgtable.h 	return __pte((pte_val(pte) & _PAGE_CHG_MASK) | pgprot_val(newprot));
pte               526 arch/powerpc/include/asm/book3s/32/pgtable.h 				pte_t *ptep, pte_t pte, int percpu)
pte               537 arch/powerpc/include/asm/book3s/32/pgtable.h 			      | (pte_val(pte) & ~_PAGE_HASHPTE));
pte               539 arch/powerpc/include/asm/book3s/32/pgtable.h 		pte_update(ptep, ~_PAGE_HASHPTE, pte_val(pte));
pte               552 arch/powerpc/include/asm/book3s/32/pgtable.h 			      | (pte_val(pte) & ~_PAGE_HASHPTE));
pte               562 arch/powerpc/include/asm/book3s/32/pgtable.h 	: "r" (pte) : "memory");
pte               571 arch/powerpc/include/asm/book3s/32/pgtable.h 		      | (pte_val(pte) & ~_PAGE_HASHPTE));
pte                79 arch/powerpc/include/asm/book3s/64/hash-64k.h static inline real_pte_t __real_pte(pte_t pte, pte_t *ptep, int offset)
pte                84 arch/powerpc/include/asm/book3s/64/hash-64k.h 	rpte.pte = pte;
pte               139 arch/powerpc/include/asm/book3s/64/hash-64k.h #define __rpte_to_pte(r)	((r).pte)
pte               157 arch/powerpc/include/asm/book3s/64/hash-64k.h #define pte_pagesize_index(mm, addr, pte)	\
pte               158 arch/powerpc/include/asm/book3s/64/hash-64k.h 	(((pte) & H_PAGE_COMBO)? MMU_PAGE_4K: MMU_PAGE_64K)
pte               147 arch/powerpc/include/asm/book3s/64/hash.h 			    pte_t *ptep, unsigned long pte, int huge);
pte               211 arch/powerpc/include/asm/book3s/64/hash.h static inline int hash__pte_none(pte_t pte)
pte               213 arch/powerpc/include/asm/book3s/64/hash.h 	return (pte_val(pte) & ~H_PTE_NONE_MASK) == 0;
pte               225 arch/powerpc/include/asm/book3s/64/hash.h 				  pte_t *ptep, pte_t pte, int percpu)
pte               231 arch/powerpc/include/asm/book3s/64/hash.h 	*ptep = pte;
pte                18 arch/powerpc/include/asm/book3s/64/hugetlb.h 						pte_t old_pte, pte_t pte);
pte               463 arch/powerpc/include/asm/book3s/64/mmu-hash.h unsigned int hash_page_do_lazy_icache(unsigned int pp, pte_t pte, int trap);
pte               489 arch/powerpc/include/asm/book3s/64/mmu-hash.h 			       unsigned long pte);
pte               153 arch/powerpc/include/asm/book3s/64/pgalloc.h 				       pte_t *pte)
pte               155 arch/powerpc/include/asm/book3s/64/pgalloc.h 	*pmd = __pmd(__pgtable_ptr_val(pte) | PMD_VAL_BITS);
pte               335 arch/powerpc/include/asm/book3s/64/pgtable.h #define __rpte_to_pte(r)	((r).pte)
pte               349 arch/powerpc/include/asm/book3s/64/pgtable.h #define pte_pagesize_index(mm, addr, pte)	MMU_PAGE_4K
pte               390 arch/powerpc/include/asm/book3s/64/pgtable.h static inline int __pte_write(pte_t pte)
pte               392 arch/powerpc/include/asm/book3s/64/pgtable.h 	return !!(pte_raw(pte) & cpu_to_be64(_PAGE_WRITE));
pte               397 arch/powerpc/include/asm/book3s/64/pgtable.h static inline bool pte_savedwrite(pte_t pte)
pte               406 arch/powerpc/include/asm/book3s/64/pgtable.h 	return !(pte_raw(pte) & cpu_to_be64(_PAGE_RWX | _PAGE_PRIVILEGED));
pte               410 arch/powerpc/include/asm/book3s/64/pgtable.h static inline bool pte_savedwrite(pte_t pte)
pte               416 arch/powerpc/include/asm/book3s/64/pgtable.h static inline int pte_write(pte_t pte)
pte               418 arch/powerpc/include/asm/book3s/64/pgtable.h 	return __pte_write(pte) || pte_savedwrite(pte);
pte               421 arch/powerpc/include/asm/book3s/64/pgtable.h static inline int pte_read(pte_t pte)
pte               423 arch/powerpc/include/asm/book3s/64/pgtable.h 	return !!(pte_raw(pte) & cpu_to_be64(_PAGE_READ));
pte               480 arch/powerpc/include/asm/book3s/64/pgtable.h static inline int pte_dirty(pte_t pte)
pte               482 arch/powerpc/include/asm/book3s/64/pgtable.h 	return !!(pte_raw(pte) & cpu_to_be64(_PAGE_DIRTY));
pte               485 arch/powerpc/include/asm/book3s/64/pgtable.h static inline int pte_young(pte_t pte)
pte               487 arch/powerpc/include/asm/book3s/64/pgtable.h 	return !!(pte_raw(pte) & cpu_to_be64(_PAGE_ACCESSED));
pte               490 arch/powerpc/include/asm/book3s/64/pgtable.h static inline int pte_special(pte_t pte)
pte               492 arch/powerpc/include/asm/book3s/64/pgtable.h 	return !!(pte_raw(pte) & cpu_to_be64(_PAGE_SPECIAL));
pte               495 arch/powerpc/include/asm/book3s/64/pgtable.h static inline bool pte_exec(pte_t pte)
pte               497 arch/powerpc/include/asm/book3s/64/pgtable.h 	return !!(pte_raw(pte) & cpu_to_be64(_PAGE_EXEC));
pte               502 arch/powerpc/include/asm/book3s/64/pgtable.h static inline bool pte_soft_dirty(pte_t pte)
pte               504 arch/powerpc/include/asm/book3s/64/pgtable.h 	return !!(pte_raw(pte) & cpu_to_be64(_PAGE_SOFT_DIRTY));
pte               507 arch/powerpc/include/asm/book3s/64/pgtable.h static inline pte_t pte_mksoft_dirty(pte_t pte)
pte               509 arch/powerpc/include/asm/book3s/64/pgtable.h 	return __pte_raw(pte_raw(pte) | cpu_to_be64(_PAGE_SOFT_DIRTY));
pte               512 arch/powerpc/include/asm/book3s/64/pgtable.h static inline pte_t pte_clear_soft_dirty(pte_t pte)
pte               514 arch/powerpc/include/asm/book3s/64/pgtable.h 	return __pte_raw(pte_raw(pte) & cpu_to_be64(~_PAGE_SOFT_DIRTY));
pte               519 arch/powerpc/include/asm/book3s/64/pgtable.h static inline int pte_protnone(pte_t pte)
pte               521 arch/powerpc/include/asm/book3s/64/pgtable.h 	return (pte_raw(pte) & cpu_to_be64(_PAGE_PRESENT | _PAGE_PTE | _PAGE_RWX)) ==
pte               526 arch/powerpc/include/asm/book3s/64/pgtable.h static inline pte_t pte_mk_savedwrite(pte_t pte)
pte               533 arch/powerpc/include/asm/book3s/64/pgtable.h 	VM_BUG_ON((pte_raw(pte) & cpu_to_be64(_PAGE_PRESENT | _PAGE_RWX | _PAGE_PRIVILEGED)) !=
pte               535 arch/powerpc/include/asm/book3s/64/pgtable.h 	return __pte_raw(pte_raw(pte) & cpu_to_be64(~_PAGE_PRIVILEGED));
pte               539 arch/powerpc/include/asm/book3s/64/pgtable.h static inline pte_t pte_clear_savedwrite(pte_t pte)
pte               544 arch/powerpc/include/asm/book3s/64/pgtable.h 	VM_BUG_ON(!pte_protnone(pte));
pte               545 arch/powerpc/include/asm/book3s/64/pgtable.h 	return __pte_raw(pte_raw(pte) | cpu_to_be64(_PAGE_PRIVILEGED));
pte               549 arch/powerpc/include/asm/book3s/64/pgtable.h static inline pte_t pte_clear_savedwrite(pte_t pte)
pte               552 arch/powerpc/include/asm/book3s/64/pgtable.h 	return __pte_raw(pte_raw(pte) & cpu_to_be64(~_PAGE_WRITE));
pte               556 arch/powerpc/include/asm/book3s/64/pgtable.h static inline int pte_present(pte_t pte)
pte               564 arch/powerpc/include/asm/book3s/64/pgtable.h 	return !!(pte_raw(pte) & cpu_to_be64(_PAGE_PRESENT | _PAGE_INVALID));
pte               567 arch/powerpc/include/asm/book3s/64/pgtable.h static inline bool pte_hw_valid(pte_t pte)
pte               569 arch/powerpc/include/asm/book3s/64/pgtable.h 	return !!(pte_raw(pte) & cpu_to_be64(_PAGE_PRESENT));
pte               573 arch/powerpc/include/asm/book3s/64/pgtable.h extern bool arch_pte_access_permitted(u64 pte, bool write, bool execute);
pte               575 arch/powerpc/include/asm/book3s/64/pgtable.h static inline bool arch_pte_access_permitted(u64 pte, bool write, bool execute)
pte               581 arch/powerpc/include/asm/book3s/64/pgtable.h static inline bool pte_user(pte_t pte)
pte               583 arch/powerpc/include/asm/book3s/64/pgtable.h 	return !(pte_raw(pte) & cpu_to_be64(_PAGE_PRIVILEGED));
pte               587 arch/powerpc/include/asm/book3s/64/pgtable.h static inline bool pte_access_permitted(pte_t pte, bool write)
pte               593 arch/powerpc/include/asm/book3s/64/pgtable.h 	if (!pte_present(pte) || !pte_user(pte) || !pte_read(pte))
pte               596 arch/powerpc/include/asm/book3s/64/pgtable.h 	if (write && !pte_write(pte))
pte               599 arch/powerpc/include/asm/book3s/64/pgtable.h 	return arch_pte_access_permitted(pte_val(pte), write, 0);
pte               617 arch/powerpc/include/asm/book3s/64/pgtable.h static inline unsigned long pte_pfn(pte_t pte)
pte               619 arch/powerpc/include/asm/book3s/64/pgtable.h 	return (pte_val(pte) & PTE_RPN_MASK) >> PAGE_SHIFT;
pte               623 arch/powerpc/include/asm/book3s/64/pgtable.h static inline pte_t pte_wrprotect(pte_t pte)
pte               625 arch/powerpc/include/asm/book3s/64/pgtable.h 	if (unlikely(pte_savedwrite(pte)))
pte               626 arch/powerpc/include/asm/book3s/64/pgtable.h 		return pte_clear_savedwrite(pte);
pte               627 arch/powerpc/include/asm/book3s/64/pgtable.h 	return __pte_raw(pte_raw(pte) & cpu_to_be64(~_PAGE_WRITE));
pte               630 arch/powerpc/include/asm/book3s/64/pgtable.h static inline pte_t pte_exprotect(pte_t pte)
pte               632 arch/powerpc/include/asm/book3s/64/pgtable.h 	return __pte_raw(pte_raw(pte) & cpu_to_be64(~_PAGE_EXEC));
pte               635 arch/powerpc/include/asm/book3s/64/pgtable.h static inline pte_t pte_mkclean(pte_t pte)
pte               637 arch/powerpc/include/asm/book3s/64/pgtable.h 	return __pte_raw(pte_raw(pte) & cpu_to_be64(~_PAGE_DIRTY));
pte               640 arch/powerpc/include/asm/book3s/64/pgtable.h static inline pte_t pte_mkold(pte_t pte)
pte               642 arch/powerpc/include/asm/book3s/64/pgtable.h 	return __pte_raw(pte_raw(pte) & cpu_to_be64(~_PAGE_ACCESSED));
pte               645 arch/powerpc/include/asm/book3s/64/pgtable.h static inline pte_t pte_mkexec(pte_t pte)
pte               647 arch/powerpc/include/asm/book3s/64/pgtable.h 	return __pte_raw(pte_raw(pte) | cpu_to_be64(_PAGE_EXEC));
pte               650 arch/powerpc/include/asm/book3s/64/pgtable.h static inline pte_t pte_mkpte(pte_t pte)
pte               652 arch/powerpc/include/asm/book3s/64/pgtable.h 	return __pte_raw(pte_raw(pte) | cpu_to_be64(_PAGE_PTE));
pte               655 arch/powerpc/include/asm/book3s/64/pgtable.h static inline pte_t pte_mkwrite(pte_t pte)
pte               660 arch/powerpc/include/asm/book3s/64/pgtable.h 	return __pte_raw(pte_raw(pte) | cpu_to_be64(_PAGE_RW));
pte               663 arch/powerpc/include/asm/book3s/64/pgtable.h static inline pte_t pte_mkdirty(pte_t pte)
pte               665 arch/powerpc/include/asm/book3s/64/pgtable.h 	return __pte_raw(pte_raw(pte) | cpu_to_be64(_PAGE_DIRTY | _PAGE_SOFT_DIRTY));
pte               668 arch/powerpc/include/asm/book3s/64/pgtable.h static inline pte_t pte_mkyoung(pte_t pte)
pte               670 arch/powerpc/include/asm/book3s/64/pgtable.h 	return __pte_raw(pte_raw(pte) | cpu_to_be64(_PAGE_ACCESSED));
pte               673 arch/powerpc/include/asm/book3s/64/pgtable.h static inline pte_t pte_mkspecial(pte_t pte)
pte               675 arch/powerpc/include/asm/book3s/64/pgtable.h 	return __pte_raw(pte_raw(pte) | cpu_to_be64(_PAGE_SPECIAL));
pte               678 arch/powerpc/include/asm/book3s/64/pgtable.h static inline pte_t pte_mkhuge(pte_t pte)
pte               680 arch/powerpc/include/asm/book3s/64/pgtable.h 	return pte;
pte               683 arch/powerpc/include/asm/book3s/64/pgtable.h static inline pte_t pte_mkdevmap(pte_t pte)
pte               685 arch/powerpc/include/asm/book3s/64/pgtable.h 	return __pte_raw(pte_raw(pte) | cpu_to_be64(_PAGE_SPECIAL | _PAGE_DEVMAP));
pte               688 arch/powerpc/include/asm/book3s/64/pgtable.h static inline pte_t pte_mkprivileged(pte_t pte)
pte               690 arch/powerpc/include/asm/book3s/64/pgtable.h 	return __pte_raw(pte_raw(pte) | cpu_to_be64(_PAGE_PRIVILEGED));
pte               693 arch/powerpc/include/asm/book3s/64/pgtable.h static inline pte_t pte_mkuser(pte_t pte)
pte               695 arch/powerpc/include/asm/book3s/64/pgtable.h 	return __pte_raw(pte_raw(pte) & cpu_to_be64(~_PAGE_PRIVILEGED));
pte               704 arch/powerpc/include/asm/book3s/64/pgtable.h static inline int pte_devmap(pte_t pte)
pte               708 arch/powerpc/include/asm/book3s/64/pgtable.h 	return (pte_raw(pte) & mask) == mask;
pte               711 arch/powerpc/include/asm/book3s/64/pgtable.h static inline pte_t pte_modify(pte_t pte, pgprot_t newprot)
pte               714 arch/powerpc/include/asm/book3s/64/pgtable.h 	return __pte_raw((pte_raw(pte) & cpu_to_be64(_PAGE_CHG_MASK)) |
pte               742 arch/powerpc/include/asm/book3s/64/pgtable.h #define __pte_to_swp_entry(pte)	((swp_entry_t) { pte_val((pte)) & ~_PAGE_PTE })
pte               754 arch/powerpc/include/asm/book3s/64/pgtable.h static inline pte_t pte_swp_mksoft_dirty(pte_t pte)
pte               756 arch/powerpc/include/asm/book3s/64/pgtable.h 	return __pte_raw(pte_raw(pte) | cpu_to_be64(_PAGE_SWP_SOFT_DIRTY));
pte               759 arch/powerpc/include/asm/book3s/64/pgtable.h static inline bool pte_swp_soft_dirty(pte_t pte)
pte               761 arch/powerpc/include/asm/book3s/64/pgtable.h 	return !!(pte_raw(pte) & cpu_to_be64(_PAGE_SWP_SOFT_DIRTY));
pte               764 arch/powerpc/include/asm/book3s/64/pgtable.h static inline pte_t pte_swp_clear_soft_dirty(pte_t pte)
pte               766 arch/powerpc/include/asm/book3s/64/pgtable.h 	return __pte_raw(pte_raw(pte) & cpu_to_be64(~_PAGE_SWP_SOFT_DIRTY));
pte               808 arch/powerpc/include/asm/book3s/64/pgtable.h static inline int pte_none(pte_t pte)
pte               811 arch/powerpc/include/asm/book3s/64/pgtable.h 		return radix__pte_none(pte);
pte               812 arch/powerpc/include/asm/book3s/64/pgtable.h 	return hash__pte_none(pte);
pte               816 arch/powerpc/include/asm/book3s/64/pgtable.h 				pte_t *ptep, pte_t pte, int percpu)
pte               819 arch/powerpc/include/asm/book3s/64/pgtable.h 		return radix__set_pte_at(mm, addr, ptep, pte, percpu);
pte               820 arch/powerpc/include/asm/book3s/64/pgtable.h 	return hash__set_pte_at(mm, addr, ptep, pte, percpu);
pte               853 arch/powerpc/include/asm/book3s/64/pgtable.h static inline bool pte_ci(pte_t pte)
pte               855 arch/powerpc/include/asm/book3s/64/pgtable.h 	__be64 pte_v = pte_raw(pte);
pte               933 arch/powerpc/include/asm/book3s/64/pgtable.h static inline pud_t pte_pud(pte_t pte)
pte               935 arch/powerpc/include/asm/book3s/64/pgtable.h 	return __pud_raw(pte_raw(pte));
pte               974 arch/powerpc/include/asm/book3s/64/pgtable.h static inline pgd_t pte_pgd(pte_t pte)
pte               976 arch/powerpc/include/asm/book3s/64/pgtable.h 	return __pgd_raw(pte_raw(pte));
pte              1022 arch/powerpc/include/asm/book3s/64/pgtable.h static inline void pte_unmap(pte_t *pte) { }
pte              1073 arch/powerpc/include/asm/book3s/64/pgtable.h static inline pmd_t pte_pmd(pte_t pte)
pte              1075 arch/powerpc/include/asm/book3s/64/pgtable.h 	return __pmd_raw(pte_raw(pte));
pte               130 arch/powerpc/include/asm/book3s/64/radix.h 					   pte_t old_pte, pte_t pte);
pte               185 arch/powerpc/include/asm/book3s/64/radix.h static inline int radix__pte_none(pte_t pte)
pte               187 arch/powerpc/include/asm/book3s/64/radix.h 	return (pte_val(pte) & ~RADIX_PTE_NONE_MASK) == 0;
pte               191 arch/powerpc/include/asm/book3s/64/radix.h 				 pte_t *ptep, pte_t pte, int percpu)
pte               193 arch/powerpc/include/asm/book3s/64/radix.h 	*ptep = pte;
pte                18 arch/powerpc/include/asm/book3s/64/tlbflush-hash.h 	real_pte_t		pte[PPC64_TLB_BATCH_NR];
pte                56 arch/powerpc/include/asm/book3s/64/tlbflush-hash.h extern void flush_hash_page(unsigned long vpn, real_pte_t pte, int psize,
pte                17 arch/powerpc/include/asm/book3s/pgtable.h 		       pte_t pte);
pte                61 arch/powerpc/include/asm/hugetlb.h 			       pte_t pte, int dirty);
pte                57 arch/powerpc/include/asm/kvm_book3s.h 	struct kvmppc_pte pte;
pte               152 arch/powerpc/include/asm/kvm_book3s.h extern int kvmppc_mmu_map_page(struct kvm_vcpu *vcpu, struct kvmppc_pte *pte,
pte               154 arch/powerpc/include/asm/kvm_book3s.h extern void kvmppc_mmu_unmap_page(struct kvm_vcpu *vcpu, struct kvmppc_pte *pte);
pte               166 arch/powerpc/include/asm/kvm_book3s.h extern void kvmppc_mmu_hpte_cache_map(struct kvm_vcpu *vcpu, struct hpte_cache *pte);
pte               168 arch/powerpc/include/asm/kvm_book3s.h extern void kvmppc_mmu_hpte_cache_free(struct hpte_cache *pte);
pte               171 arch/powerpc/include/asm/kvm_book3s.h extern void kvmppc_mmu_invalidate_pte(struct kvm_vcpu *vcpu, struct hpte_cache *pte);
pte               197 arch/powerpc/include/asm/kvm_book3s.h extern void kvmppc_unmap_pte(struct kvm *kvm, pte_t *pte, unsigned long gpa,
pte               623 arch/powerpc/include/asm/kvm_book3s_64.h extern int kvmppc_create_pte(struct kvm *kvm, pgd_t *pgtable, pte_t pte,
pte                64 arch/powerpc/include/asm/kvm_host.h extern int kvm_set_spte_hva(struct kvm *kvm, unsigned long hva, pte_t pte);
pte               403 arch/powerpc/include/asm/kvm_host.h 		      struct kvmppc_pte *pte, bool data, bool iswrite);
pte               120 arch/powerpc/include/asm/kvm_ppc.h 			struct kvmppc_pte *pte);
pte               293 arch/powerpc/include/asm/kvm_ppc.h 	void (*set_spte_hva)(struct kvm *kvm, unsigned long hva, pte_t pte);
pte                18 arch/powerpc/include/asm/nohash/32/pgalloc.h 				       pte_t *pte)
pte                21 arch/powerpc/include/asm/nohash/32/pgalloc.h 		*pmdp = __pmd((unsigned long)pte | _PMD_PRESENT);
pte                23 arch/powerpc/include/asm/nohash/32/pgalloc.h 		*pmdp = __pmd(__pa(pte) | _PMD_PRESENT);
pte               167 arch/powerpc/include/asm/nohash/32/pgtable.h static inline pte_t pte_mkwrite(pte_t pte)
pte               169 arch/powerpc/include/asm/nohash/32/pgtable.h 	return __pte(pte_val(pte) | _PAGE_RW);
pte               173 arch/powerpc/include/asm/nohash/32/pgtable.h static inline pte_t pte_mkdirty(pte_t pte)
pte               175 arch/powerpc/include/asm/nohash/32/pgtable.h 	return __pte(pte_val(pte) | _PAGE_DIRTY);
pte               178 arch/powerpc/include/asm/nohash/32/pgtable.h static inline pte_t pte_mkyoung(pte_t pte)
pte               180 arch/powerpc/include/asm/nohash/32/pgtable.h 	return __pte(pte_val(pte) | _PAGE_ACCESSED);
pte               184 arch/powerpc/include/asm/nohash/32/pgtable.h static inline pte_t pte_wrprotect(pte_t pte)
pte               186 arch/powerpc/include/asm/nohash/32/pgtable.h 	return __pte(pte_val(pte) & ~_PAGE_RW);
pte               190 arch/powerpc/include/asm/nohash/32/pgtable.h static inline pte_t pte_mkexec(pte_t pte)
pte               192 arch/powerpc/include/asm/nohash/32/pgtable.h 	return __pte(pte_val(pte) | _PAGE_EXEC);
pte               243 arch/powerpc/include/asm/nohash/32/pgtable.h 	p->pte = p->pte1 = p->pte2 = p->pte3 = new;
pte               330 arch/powerpc/include/asm/nohash/32/pgtable.h static inline int pte_young(pte_t pte)
pte               332 arch/powerpc/include/asm/nohash/32/pgtable.h 	return pte_val(pte) & _PAGE_ACCESSED;
pte               373 arch/powerpc/include/asm/nohash/32/pgtable.h #define pte_unmap(pte)		kunmap_atomic(pte)
pte               384 arch/powerpc/include/asm/nohash/32/pgtable.h #define __pte_to_swp_entry(pte)		((swp_entry_t) { pte_val(pte) >> 3 })
pte                89 arch/powerpc/include/asm/nohash/32/pte-40x.h static inline pte_t pte_wrprotect(pte_t pte)
pte                91 arch/powerpc/include/asm/nohash/32/pte-40x.h 	return __pte(pte_val(pte) & ~(_PAGE_RW | _PAGE_HWWRITE));
pte                96 arch/powerpc/include/asm/nohash/32/pte-40x.h static inline pte_t pte_mkclean(pte_t pte)
pte                98 arch/powerpc/include/asm/nohash/32/pte-40x.h 	return __pte(pte_val(pte) & ~(_PAGE_DIRTY | _PAGE_HWWRITE));
pte                87 arch/powerpc/include/asm/nohash/32/pte-8xx.h static inline pte_t pte_wrprotect(pte_t pte)
pte                89 arch/powerpc/include/asm/nohash/32/pte-8xx.h 	return __pte(pte_val(pte) | _PAGE_RO);
pte                94 arch/powerpc/include/asm/nohash/32/pte-8xx.h static inline int pte_write(pte_t pte)
pte                96 arch/powerpc/include/asm/nohash/32/pte-8xx.h 	return !(pte_val(pte) & _PAGE_RO);
pte               101 arch/powerpc/include/asm/nohash/32/pte-8xx.h static inline pte_t pte_mkwrite(pte_t pte)
pte               103 arch/powerpc/include/asm/nohash/32/pte-8xx.h 	return __pte(pte_val(pte) & ~_PAGE_RO);
pte               108 arch/powerpc/include/asm/nohash/32/pte-8xx.h static inline bool pte_user(pte_t pte)
pte               110 arch/powerpc/include/asm/nohash/32/pte-8xx.h 	return !(pte_val(pte) & _PAGE_SH);
pte               115 arch/powerpc/include/asm/nohash/32/pte-8xx.h static inline pte_t pte_mkprivileged(pte_t pte)
pte               117 arch/powerpc/include/asm/nohash/32/pte-8xx.h 	return __pte(pte_val(pte) | _PAGE_SH);
pte               122 arch/powerpc/include/asm/nohash/32/pte-8xx.h static inline pte_t pte_mkuser(pte_t pte)
pte               124 arch/powerpc/include/asm/nohash/32/pte-8xx.h 	return __pte(pte_val(pte) & ~_PAGE_SH);
pte               129 arch/powerpc/include/asm/nohash/32/pte-8xx.h static inline pte_t pte_mkhuge(pte_t pte)
pte               131 arch/powerpc/include/asm/nohash/32/pte-8xx.h 	return __pte(pte_val(pte) | _PAGE_SPS);
pte                37 arch/powerpc/include/asm/nohash/64/pgalloc.h 				       pte_t *pte)
pte                39 arch/powerpc/include/asm/nohash/64/pgalloc.h 	pmd_set(pmd, (unsigned long)pte);
pte                73 arch/powerpc/include/asm/nohash/64/pgtable-4k.h static inline pgd_t pte_pgd(pte_t pte)
pte                75 arch/powerpc/include/asm/nohash/64/pgtable-4k.h 	return __pgd(pte_val(pte));
pte               100 arch/powerpc/include/asm/nohash/64/pgtable.h static inline pte_t pte_mkwrite(pte_t pte)
pte               102 arch/powerpc/include/asm/nohash/64/pgtable.h 	return __pte(pte_val(pte) | _PAGE_RW);
pte               105 arch/powerpc/include/asm/nohash/64/pgtable.h static inline pte_t pte_mkdirty(pte_t pte)
pte               107 arch/powerpc/include/asm/nohash/64/pgtable.h 	return __pte(pte_val(pte) | _PAGE_DIRTY);
pte               110 arch/powerpc/include/asm/nohash/64/pgtable.h static inline pte_t pte_mkyoung(pte_t pte)
pte               112 arch/powerpc/include/asm/nohash/64/pgtable.h 	return __pte(pte_val(pte) | _PAGE_ACCESSED);
pte               115 arch/powerpc/include/asm/nohash/64/pgtable.h static inline pte_t pte_wrprotect(pte_t pte)
pte               117 arch/powerpc/include/asm/nohash/64/pgtable.h 	return __pte(pte_val(pte) & ~_PAGE_RW);
pte               120 arch/powerpc/include/asm/nohash/64/pgtable.h static inline pte_t pte_mkexec(pte_t pte)
pte               122 arch/powerpc/include/asm/nohash/64/pgtable.h 	return __pte(pte_val(pte) | _PAGE_EXEC);
pte               173 arch/powerpc/include/asm/nohash/64/pgtable.h static inline pud_t pte_pud(pte_t pte)
pte               175 arch/powerpc/include/asm/nohash/64/pgtable.h 	return __pud(pte_val(pte));
pte               201 arch/powerpc/include/asm/nohash/64/pgtable.h static inline void pte_unmap(pte_t *pte) { }
pte               237 arch/powerpc/include/asm/nohash/64/pgtable.h static inline int pte_young(pte_t pte)
pte               239 arch/powerpc/include/asm/nohash/64/pgtable.h 	return pte_val(pte) & _PAGE_ACCESSED;
pte               355 arch/powerpc/include/asm/nohash/64/pgtable.h #define __pte_to_swp_entry(pte)		((swp_entry_t) { pte_val((pte)) })
pte                43 arch/powerpc/include/asm/nohash/pgtable.h static inline int pte_write(pte_t pte)
pte                45 arch/powerpc/include/asm/nohash/pgtable.h 	return pte_val(pte) & _PAGE_RW;
pte                48 arch/powerpc/include/asm/nohash/pgtable.h static inline int pte_read(pte_t pte)		{ return 1; }
pte                49 arch/powerpc/include/asm/nohash/pgtable.h static inline int pte_dirty(pte_t pte)		{ return pte_val(pte) & _PAGE_DIRTY; }
pte                50 arch/powerpc/include/asm/nohash/pgtable.h static inline int pte_special(pte_t pte)	{ return pte_val(pte) & _PAGE_SPECIAL; }
pte                51 arch/powerpc/include/asm/nohash/pgtable.h static inline int pte_none(pte_t pte)		{ return (pte_val(pte) & ~_PTE_NONE_MASK) == 0; }
pte                52 arch/powerpc/include/asm/nohash/pgtable.h static inline bool pte_hashpte(pte_t pte)	{ return false; }
pte                53 arch/powerpc/include/asm/nohash/pgtable.h static inline bool pte_ci(pte_t pte)		{ return pte_val(pte) & _PAGE_NO_CACHE; }
pte                54 arch/powerpc/include/asm/nohash/pgtable.h static inline bool pte_exec(pte_t pte)		{ return pte_val(pte) & _PAGE_EXEC; }
pte                62 arch/powerpc/include/asm/nohash/pgtable.h static inline int pte_protnone(pte_t pte)
pte                64 arch/powerpc/include/asm/nohash/pgtable.h 	return pte_present(pte) && !pte_user(pte);
pte                73 arch/powerpc/include/asm/nohash/pgtable.h static inline int pte_present(pte_t pte)
pte                75 arch/powerpc/include/asm/nohash/pgtable.h 	return pte_val(pte) & _PAGE_PRESENT;
pte                78 arch/powerpc/include/asm/nohash/pgtable.h static inline bool pte_hw_valid(pte_t pte)
pte                80 arch/powerpc/include/asm/nohash/pgtable.h 	return pte_val(pte) & _PAGE_PRESENT;
pte                89 arch/powerpc/include/asm/nohash/pgtable.h static inline bool pte_user(pte_t pte)
pte                91 arch/powerpc/include/asm/nohash/pgtable.h 	return (pte_val(pte) & _PAGE_USER) == _PAGE_USER;
pte               100 arch/powerpc/include/asm/nohash/pgtable.h static inline bool pte_access_permitted(pte_t pte, bool write)
pte               106 arch/powerpc/include/asm/nohash/pgtable.h 	if (!pte_present(pte) || !pte_user(pte) || !pte_read(pte))
pte               109 arch/powerpc/include/asm/nohash/pgtable.h 	if (write && !pte_write(pte))
pte               124 arch/powerpc/include/asm/nohash/pgtable.h static inline unsigned long pte_pfn(pte_t pte)	{
pte               125 arch/powerpc/include/asm/nohash/pgtable.h 	return pte_val(pte) >> PTE_RPN_SHIFT; }
pte               128 arch/powerpc/include/asm/nohash/pgtable.h static inline pte_t pte_exprotect(pte_t pte)
pte               130 arch/powerpc/include/asm/nohash/pgtable.h 	return __pte(pte_val(pte) & ~_PAGE_EXEC);
pte               134 arch/powerpc/include/asm/nohash/pgtable.h static inline pte_t pte_mkclean(pte_t pte)
pte               136 arch/powerpc/include/asm/nohash/pgtable.h 	return __pte(pte_val(pte) & ~_PAGE_DIRTY);
pte               140 arch/powerpc/include/asm/nohash/pgtable.h static inline pte_t pte_mkold(pte_t pte)
pte               142 arch/powerpc/include/asm/nohash/pgtable.h 	return __pte(pte_val(pte) & ~_PAGE_ACCESSED);
pte               145 arch/powerpc/include/asm/nohash/pgtable.h static inline pte_t pte_mkpte(pte_t pte)
pte               147 arch/powerpc/include/asm/nohash/pgtable.h 	return pte;
pte               150 arch/powerpc/include/asm/nohash/pgtable.h static inline pte_t pte_mkspecial(pte_t pte)
pte               152 arch/powerpc/include/asm/nohash/pgtable.h 	return __pte(pte_val(pte) | _PAGE_SPECIAL);
pte               156 arch/powerpc/include/asm/nohash/pgtable.h static inline pte_t pte_mkhuge(pte_t pte)
pte               158 arch/powerpc/include/asm/nohash/pgtable.h 	return __pte(pte_val(pte));
pte               163 arch/powerpc/include/asm/nohash/pgtable.h static inline pte_t pte_mkprivileged(pte_t pte)
pte               165 arch/powerpc/include/asm/nohash/pgtable.h 	return __pte(pte_val(pte) & ~_PAGE_USER);
pte               170 arch/powerpc/include/asm/nohash/pgtable.h static inline pte_t pte_mkuser(pte_t pte)
pte               172 arch/powerpc/include/asm/nohash/pgtable.h 	return __pte(pte_val(pte) | _PAGE_USER);
pte               176 arch/powerpc/include/asm/nohash/pgtable.h static inline pte_t pte_modify(pte_t pte, pgprot_t newprot)
pte               178 arch/powerpc/include/asm/nohash/pgtable.h 	return __pte((pte_val(pte) & _PAGE_CHG_MASK) | pgprot_val(newprot));
pte               185 arch/powerpc/include/asm/nohash/pgtable.h 		       pte_t pte);
pte               193 arch/powerpc/include/asm/nohash/pgtable.h 				pte_t *ptep, pte_t pte, int percpu)
pte               206 arch/powerpc/include/asm/nohash/pgtable.h 		: "r" (pte) : "memory");
pte               213 arch/powerpc/include/asm/nohash/pgtable.h 	ptep->pte = ptep->pte1 = ptep->pte2 = ptep->pte3 = pte_val(pte);
pte               215 arch/powerpc/include/asm/nohash/pgtable.h 	*ptep = pte;
pte               103 arch/powerpc/include/asm/nohash/pte-book3e.h static inline pte_t pte_mkprivileged(pte_t pte)
pte               105 arch/powerpc/include/asm/nohash/pte-book3e.h 	return __pte((pte_val(pte) & ~_PAGE_USER) | _PAGE_PRIVILEGED);
pte               110 arch/powerpc/include/asm/nohash/pte-book3e.h static inline pte_t pte_mkuser(pte_t pte)
pte               112 arch/powerpc/include/asm/nohash/pte-book3e.h 	return __pte((pte_val(pte) & ~_PAGE_PRIVILEGED) | _PAGE_USER);
pte                38 arch/powerpc/include/asm/pgalloc.h static inline void pte_free_kernel(struct mm_struct *mm, pte_t *pte)
pte                40 arch/powerpc/include/asm/pgalloc.h 	pte_fragment_free((unsigned long *)pte, 1);
pte                 8 arch/powerpc/include/asm/pgtable-be-types.h typedef struct { __be64 pte; } pte_t;
pte                13 arch/powerpc/include/asm/pgtable-be-types.h 	return be64_to_cpu(x.pte);
pte                18 arch/powerpc/include/asm/pgtable-be-types.h 	return x.pte;
pte                76 arch/powerpc/include/asm/pgtable-be-types.h typedef struct { pte_t pte; unsigned long hidx; } real_pte_t;
pte                78 arch/powerpc/include/asm/pgtable-be-types.h typedef struct { pte_t pte; } real_pte_t;
pte                 7 arch/powerpc/include/asm/pgtable-types.h typedef struct { pte_basic_t pte, pte1, pte2, pte3; } pte_t;
pte                 9 arch/powerpc/include/asm/pgtable-types.h typedef struct { pte_basic_t pte; } pte_t;
pte                14 arch/powerpc/include/asm/pgtable-types.h 	return x.pte;
pte                53 arch/powerpc/include/asm/pgtable-types.h typedef struct { pte_t pte; unsigned long hidx; } real_pte_t;
pte                55 arch/powerpc/include/asm/pgtable-types.h typedef struct { pte_t pte; } real_pte_t;
pte                52 arch/powerpc/include/asm/pgtable.h static inline pgprot_t pte_pgprot(pte_t pte)
pte                56 arch/powerpc/include/asm/pgtable.h 	pte_flags = pte_val(pte) & ~PTE_RPN_MASK;
pte                13 arch/powerpc/include/asm/pte-walk.h 	pte_t *pte;
pte                16 arch/powerpc/include/asm/pte-walk.h 	pte = __find_linux_pte(pgdir, ea, is_thp, hshift);
pte                26 arch/powerpc/include/asm/pte-walk.h 	return pte;
pte                41 arch/powerpc/include/asm/pte-walk.h 	pte_t *pte;
pte                46 arch/powerpc/include/asm/pte-walk.h 	pte = __find_linux_pte(pgdir, ea, is_thp, hshift);
pte                56 arch/powerpc/include/asm/pte-walk.h 	return pte;
pte               453 arch/powerpc/kvm/book3s.c 		 enum xlate_readwrite xlrw, struct kvmppc_pte *pte)
pte               461 arch/powerpc/kvm/book3s.c 		r = vcpu->arch.mmu.xlate(vcpu, eaddr, pte, data, iswrite);
pte               463 arch/powerpc/kvm/book3s.c 		pte->eaddr = eaddr;
pte               464 arch/powerpc/kvm/book3s.c 		pte->raddr = eaddr & KVM_PAM;
pte               465 arch/powerpc/kvm/book3s.c 		pte->vpage = VSID_REAL | eaddr >> 12;
pte               466 arch/powerpc/kvm/book3s.c 		pte->may_read = true;
pte               467 arch/powerpc/kvm/book3s.c 		pte->may_write = true;
pte               468 arch/powerpc/kvm/book3s.c 		pte->may_execute = true;
pte               475 arch/powerpc/kvm/book3s.c 			pte->raddr &= ~SPLIT_HACK_MASK;
pte               885 arch/powerpc/kvm/book3s.c int kvm_set_spte_hva(struct kvm *kvm, unsigned long hva, pte_t pte)
pte               887 arch/powerpc/kvm/book3s.c 	kvm->arch.kvm_ops->set_spte_hva(kvm, hva, pte);
pte                17 arch/powerpc/kvm/book3s.h extern void kvm_set_spte_hva_hv(struct kvm *kvm, unsigned long hva, pte_t pte);
pte                70 arch/powerpc/kvm/book3s_32_mmu.c 					  struct kvmppc_pte *pte, bool data,
pte                84 arch/powerpc/kvm/book3s_32_mmu.c 	struct kvmppc_pte pte;
pte                86 arch/powerpc/kvm/book3s_32_mmu.c 	if (!kvmppc_mmu_book3s_32_xlate_bat(vcpu, eaddr, &pte, data, false))
pte                87 arch/powerpc/kvm/book3s_32_mmu.c 		return pte.vpage;
pte               133 arch/powerpc/kvm/book3s_32_mmu.c 					  struct kvmppc_pte *pte, bool data,
pte               165 arch/powerpc/kvm/book3s_32_mmu.c 			pte->vpage = (((u64)eaddr >> 12) & 0xffff) | vsid;
pte               167 arch/powerpc/kvm/book3s_32_mmu.c 			pte->raddr = bat->brpn | (eaddr & ~bat->bepi_mask);
pte               168 arch/powerpc/kvm/book3s_32_mmu.c 			pte->may_read = bat->pp;
pte               169 arch/powerpc/kvm/book3s_32_mmu.c 			pte->may_write = bat->pp > 1;
pte               170 arch/powerpc/kvm/book3s_32_mmu.c 			pte->may_execute = true;
pte               171 arch/powerpc/kvm/book3s_32_mmu.c 			if (!pte->may_read) {
pte               175 arch/powerpc/kvm/book3s_32_mmu.c 			if (iswrite && !pte->may_write) {
pte               188 arch/powerpc/kvm/book3s_32_mmu.c 				     struct kvmppc_pte *pte, bool data,
pte               204 arch/powerpc/kvm/book3s_32_mmu.c 	pte->vpage = kvmppc_mmu_book3s_32_ea_to_vp(vcpu, eaddr, data);
pte               226 arch/powerpc/kvm/book3s_32_mmu.c 			pte->raddr = (pte1 & ~(0xFFFULL)) | (eaddr & 0xFFF);
pte               233 arch/powerpc/kvm/book3s_32_mmu.c 			pte->may_write = false;
pte               234 arch/powerpc/kvm/book3s_32_mmu.c 			pte->may_read = false;
pte               235 arch/powerpc/kvm/book3s_32_mmu.c 			pte->may_execute = true;
pte               241 arch/powerpc/kvm/book3s_32_mmu.c 					pte->may_write = true;
pte               246 arch/powerpc/kvm/book3s_32_mmu.c 					pte->may_read = true;
pte               267 arch/powerpc/kvm/book3s_32_mmu.c 		if (pte->may_read && !(pte_r & PTEG_FLAG_ACCESSED)) {
pte               271 arch/powerpc/kvm/book3s_32_mmu.c 		if (iswrite && pte->may_write && !(pte_r & PTEG_FLAG_DIRTY)) {
pte               275 arch/powerpc/kvm/book3s_32_mmu.c 		if (!pte->may_read || (iswrite && !pte->may_write))
pte               296 arch/powerpc/kvm/book3s_32_mmu.c 				      struct kvmppc_pte *pte, bool data,
pte               302 arch/powerpc/kvm/book3s_32_mmu.c 	pte->eaddr = eaddr;
pte               303 arch/powerpc/kvm/book3s_32_mmu.c 	pte->page_size = MMU_PAGE_4K;
pte               309 arch/powerpc/kvm/book3s_32_mmu.c 		pte->vpage = kvmppc_mmu_book3s_32_ea_to_vp(vcpu, eaddr, data);
pte               310 arch/powerpc/kvm/book3s_32_mmu.c 		pte->raddr = vcpu->arch.magic_page_pa | (pte->raddr & 0xfff);
pte               311 arch/powerpc/kvm/book3s_32_mmu.c 		pte->raddr &= KVM_PAM;
pte               312 arch/powerpc/kvm/book3s_32_mmu.c 		pte->may_execute = true;
pte               313 arch/powerpc/kvm/book3s_32_mmu.c 		pte->may_read = true;
pte               314 arch/powerpc/kvm/book3s_32_mmu.c 		pte->may_write = true;
pte               319 arch/powerpc/kvm/book3s_32_mmu.c 	r = kvmppc_mmu_book3s_32_xlate_bat(vcpu, eaddr, pte, data, iswrite);
pte               321 arch/powerpc/kvm/book3s_32_mmu.c 		r = kvmppc_mmu_book3s_32_xlate_pte(vcpu, eaddr, pte,
pte               324 arch/powerpc/kvm/book3s_32_mmu.c 		r = kvmppc_mmu_book3s_32_xlate_pte(vcpu, eaddr, pte,
pte                49 arch/powerpc/kvm/book3s_32_mmu_host.c void kvmppc_mmu_invalidate_pte(struct kvm_vcpu *vcpu, struct hpte_cache *pte)
pte                54 arch/powerpc/kvm/book3s_32_mmu_host.c 	pteg = (u32*)pte->slot;
pte                59 arch/powerpc/kvm/book3s_32_mmu_host.c 	asm volatile ("tlbie %0" : : "r" (pte->pte.eaddr) : "memory");
pte               143 arch/powerpc/kvm/book3s_32_mmu_host.c 	struct hpte_cache *pte;
pte               233 arch/powerpc/kvm/book3s_32_mmu_host.c 	pte = kvmppc_mmu_hpte_cache_next(vcpu);
pte               234 arch/powerpc/kvm/book3s_32_mmu_host.c 	if (!pte) {
pte               246 arch/powerpc/kvm/book3s_32_mmu_host.c 	pte->slot = (ulong)&pteg[rr];
pte               247 arch/powerpc/kvm/book3s_32_mmu_host.c 	pte->host_vpn = vpn;
pte               248 arch/powerpc/kvm/book3s_32_mmu_host.c 	pte->pte = *orig_pte;
pte               249 arch/powerpc/kvm/book3s_32_mmu_host.c 	pte->pfn = hpaddr >> PAGE_SHIFT;
pte               251 arch/powerpc/kvm/book3s_32_mmu_host.c 	kvmppc_mmu_hpte_cache_map(vcpu, pte);
pte               258 arch/powerpc/kvm/book3s_32_mmu_host.c void kvmppc_mmu_unmap_page(struct kvm_vcpu *vcpu, struct kvmppc_pte *pte)
pte               260 arch/powerpc/kvm/book3s_32_mmu_host.c 	kvmppc_mmu_pte_vflush(vcpu, pte->vpage, 0xfffffffffULL);
pte                23 arch/powerpc/kvm/book3s_64_mmu_host.c void kvmppc_mmu_invalidate_pte(struct kvm_vcpu *vcpu, struct hpte_cache *pte)
pte                25 arch/powerpc/kvm/book3s_64_mmu_host.c 	mmu_hash_ops.hpte_invalidate(pte->slot, pte->host_vpn,
pte                26 arch/powerpc/kvm/book3s_64_mmu_host.c 				     pte->pagesize, pte->pagesize,
pte               194 arch/powerpc/kvm/book3s_64_mmu_host.c 		cpte->pte = *orig_pte;
pte               212 arch/powerpc/kvm/book3s_64_mmu_host.c void kvmppc_mmu_unmap_page(struct kvm_vcpu *vcpu, struct kvmppc_pte *pte)
pte               217 arch/powerpc/kvm/book3s_64_mmu_host.c 	vcpu->arch.mmu.esid_to_vsid(vcpu, pte->eaddr >> SID_SHIFT, &vsid);
pte               220 arch/powerpc/kvm/book3s_64_mmu_host.c 	kvmppc_mmu_pte_vflush(vcpu, pte->vpage, mask);
pte               617 arch/powerpc/kvm/book3s_64_mmu_hv.c 			pte_t *ptep, pte;
pte               627 arch/powerpc/kvm/book3s_64_mmu_hv.c 				pte = kvmppc_read_update_linux_pte(ptep, 1);
pte               628 arch/powerpc/kvm/book3s_64_mmu_hv.c 				if (__pte_write(pte))
pte              1031 arch/powerpc/kvm/book3s_64_mmu_hv.c void kvm_set_spte_hva_hv(struct kvm *kvm, unsigned long hva, pte_t pte)
pte               134 arch/powerpc/kvm/book3s_64_mmu_radix.c 	u64 pte, base, gpa;
pte               169 arch/powerpc/kvm/book3s_64_mmu_radix.c 		pte = __be64_to_cpu(rpte);
pte               170 arch/powerpc/kvm/book3s_64_mmu_radix.c 		if (!(pte & _PAGE_PRESENT))
pte               173 arch/powerpc/kvm/book3s_64_mmu_radix.c 		if (pte & _PAGE_PTE)
pte               176 arch/powerpc/kvm/book3s_64_mmu_radix.c 		base = pte & RPDB_MASK;
pte               177 arch/powerpc/kvm/book3s_64_mmu_radix.c 		bits = pte & RPDS_MASK;
pte               186 arch/powerpc/kvm/book3s_64_mmu_radix.c 	gpa = pte & 0x01fffffffffff000ul;
pte               200 arch/powerpc/kvm/book3s_64_mmu_radix.c 	gpte->may_read = !!(pte & _PAGE_READ);
pte               201 arch/powerpc/kvm/book3s_64_mmu_radix.c 	gpte->may_write = !!(pte & _PAGE_WRITE);
pte               202 arch/powerpc/kvm/book3s_64_mmu_radix.c 	gpte->may_execute = !!(pte & _PAGE_EXEC);
pte               204 arch/powerpc/kvm/book3s_64_mmu_radix.c 	gpte->rc = pte & (_PAGE_ACCESSED | _PAGE_DIRTY);
pte               207 arch/powerpc/kvm/book3s_64_mmu_radix.c 		*pte_ret_p = pte;
pte               253 arch/powerpc/kvm/book3s_64_mmu_radix.c 	u64 pte;
pte               269 arch/powerpc/kvm/book3s_64_mmu_radix.c 				vcpu->kvm->arch.process_table, pid, &pte);
pte               275 arch/powerpc/kvm/book3s_64_mmu_radix.c 		if (pte & _PAGE_PRIVILEGED) {
pte               281 arch/powerpc/kvm/book3s_64_mmu_radix.c 		if (!(pte & _PAGE_PRIVILEGED)) {
pte               346 arch/powerpc/kvm/book3s_64_mmu_radix.c 			     pte_t *ptep, pte_t pte)
pte               348 arch/powerpc/kvm/book3s_64_mmu_radix.c 	radix__set_pte_at(kvm->mm, addr, ptep, pte, 0);
pte               375 arch/powerpc/kvm/book3s_64_mmu_radix.c void kvmppc_unmap_pte(struct kvm *kvm, pte_t *pte, unsigned long gpa,
pte               386 arch/powerpc/kvm/book3s_64_mmu_radix.c 	old = kvmppc_radix_update_pte(kvm, pte, ~0UL, 0, gpa, shift);
pte               424 arch/powerpc/kvm/book3s_64_mmu_radix.c static void kvmppc_unmap_free_pte(struct kvm *kvm, pte_t *pte, bool full,
pte               428 arch/powerpc/kvm/book3s_64_mmu_radix.c 		memset(pte, 0, sizeof(long) << PTE_INDEX_SIZE);
pte               430 arch/powerpc/kvm/book3s_64_mmu_radix.c 		pte_t *p = pte;
pte               443 arch/powerpc/kvm/book3s_64_mmu_radix.c 	kvmppc_pte_free(pte);
pte               465 arch/powerpc/kvm/book3s_64_mmu_radix.c 			pte_t *pte;
pte               467 arch/powerpc/kvm/book3s_64_mmu_radix.c 			pte = pte_offset_map(p, 0);
pte               468 arch/powerpc/kvm/book3s_64_mmu_radix.c 			kvmppc_unmap_free_pte(kvm, pte, full, lpid);
pte               525 arch/powerpc/kvm/book3s_64_mmu_radix.c 	pte_t *pte = pte_offset_kernel(pmd, 0);
pte               535 arch/powerpc/kvm/book3s_64_mmu_radix.c 	kvmppc_unmap_free_pte(kvm, pte, false, lpid);
pte               563 arch/powerpc/kvm/book3s_64_mmu_radix.c int kvmppc_create_pte(struct kvm *kvm, pgd_t *pgtable, pte_t pte,
pte               611 arch/powerpc/kvm/book3s_64_mmu_radix.c 			if (pud_raw(*pud) == pte_raw(pte)) {
pte               616 arch/powerpc/kvm/book3s_64_mmu_radix.c 			WARN_ON_ONCE((pud_val(*pud) ^ pte_val(pte)) &
pte               619 arch/powerpc/kvm/book3s_64_mmu_radix.c 					      0, pte_val(pte), hgpa, PUD_SHIFT);
pte               644 arch/powerpc/kvm/book3s_64_mmu_radix.c 		kvmppc_radix_set_pte_at(kvm, gpa, (pte_t *)pud, pte);
pte               662 arch/powerpc/kvm/book3s_64_mmu_radix.c 			if (pmd_raw(*pmd) == pte_raw(pte)) {
pte               667 arch/powerpc/kvm/book3s_64_mmu_radix.c 			WARN_ON_ONCE((pmd_val(*pmd) ^ pte_val(pte)) &
pte               670 arch/powerpc/kvm/book3s_64_mmu_radix.c 					0, pte_val(pte), lgpa, PMD_SHIFT);
pte               696 arch/powerpc/kvm/book3s_64_mmu_radix.c 		kvmppc_radix_set_pte_at(kvm, gpa, pmdp_ptep(pmd), pte);
pte               711 arch/powerpc/kvm/book3s_64_mmu_radix.c 		if (pte_raw(*ptep) == pte_raw(pte)) {
pte               716 arch/powerpc/kvm/book3s_64_mmu_radix.c 		WARN_ON_ONCE((pte_val(*ptep) ^ pte_val(pte)) &
pte               718 arch/powerpc/kvm/book3s_64_mmu_radix.c 		kvmppc_radix_update_pte(kvm, ptep, 0, pte_val(pte), gpa, 0);
pte               722 arch/powerpc/kvm/book3s_64_mmu_radix.c 	kvmppc_radix_set_pte_at(kvm, gpa, ptep, pte);
pte               778 arch/powerpc/kvm/book3s_64_mmu_radix.c 	pte_t pte, *ptep;
pte               828 arch/powerpc/kvm/book3s_64_mmu_radix.c 	pte = *ptep;
pte               852 arch/powerpc/kvm/book3s_64_mmu_radix.c 			pte = __pte(pte_val(pte) | (hva & rpnmask));
pte               856 arch/powerpc/kvm/book3s_64_mmu_radix.c 	pte = __pte(pte_val(pte) | _PAGE_EXEC | _PAGE_ACCESSED);
pte               858 arch/powerpc/kvm/book3s_64_mmu_radix.c 		if (pte_val(pte) & _PAGE_WRITE)
pte               859 arch/powerpc/kvm/book3s_64_mmu_radix.c 			pte = __pte(pte_val(pte) | _PAGE_DIRTY);
pte               861 arch/powerpc/kvm/book3s_64_mmu_radix.c 		pte = __pte(pte_val(pte) & ~(_PAGE_WRITE | _PAGE_DIRTY));
pte               865 arch/powerpc/kvm/book3s_64_mmu_radix.c 	ret = kvmppc_create_pte(kvm, kvm->arch.pgtable, pte, gpa, level,
pte               868 arch/powerpc/kvm/book3s_64_mmu_radix.c 		*inserted_pte = pte;
pte               873 arch/powerpc/kvm/book3s_64_mmu_radix.c 		if (!ret && (pte_val(pte) & _PAGE_WRITE))
pte              1204 arch/powerpc/kvm/book3s_64_mmu_radix.c 	unsigned long pte;
pte              1285 arch/powerpc/kvm/book3s_64_mmu_radix.c 			pte = pud_val(pud);
pte              1297 arch/powerpc/kvm/book3s_64_mmu_radix.c 			pte = pmd_val(pmd);
pte              1303 arch/powerpc/kvm/book3s_64_mmu_radix.c 		pte = pte_val(READ_ONCE(*ptep));
pte              1304 arch/powerpc/kvm/book3s_64_mmu_radix.c 		if (!(pte & _PAGE_PRESENT)) {
pte              1311 arch/powerpc/kvm/book3s_64_mmu_radix.c 			      " %lx: %lx %d\n", gpa, pte, shift);
pte               444 arch/powerpc/kvm/book3s_64_vio_hv.c 	pte_t *ptep, pte;
pte               459 arch/powerpc/kvm/book3s_64_vio_hv.c 	pte = *ptep;
pte               468 arch/powerpc/kvm/book3s_64_vio_hv.c 	if (!pte_young(pte))
pte               471 arch/powerpc/kvm/book3s_64_vio_hv.c 	*phpa = (pte_pfn(pte) << PAGE_SHIFT) | (ua & ((1ULL << shift) - 1)) |
pte              1268 arch/powerpc/kvm/book3s_hv_nested.c 	pte_t pte, *pte_p;
pte              1363 arch/powerpc/kvm/book3s_hv_nested.c 	pte = __pte(0);
pte              1369 arch/powerpc/kvm/book3s_hv_nested.c 		pte = *pte_p;
pte              1372 arch/powerpc/kvm/book3s_hv_nested.c 	if (!pte_present(pte) || (writing && !(pte_val(pte) & _PAGE_WRITE))) {
pte              1375 arch/powerpc/kvm/book3s_hv_nested.c 					writing, kvm_ro, &pte, &level);
pte              1394 arch/powerpc/kvm/book3s_hv_nested.c 	pte = __pte(pte_val(pte) & ~perm);
pte              1403 arch/powerpc/kvm/book3s_hv_nested.c 		pte = __pte(pte_val(pte) | (gpa & mask));
pte              1417 arch/powerpc/kvm/book3s_hv_nested.c 	ret = kvmppc_create_pte(kvm, gp->shadow_pgtable, pte, n_gpa, level,
pte               263 arch/powerpc/kvm/book3s_hv_rm_mmu.c 		pte_t pte;
pte               279 arch/powerpc/kvm/book3s_hv_rm_mmu.c 		pte = kvmppc_read_update_linux_pte(ptep, writing);
pte               280 arch/powerpc/kvm/book3s_hv_rm_mmu.c 		if (pte_present(pte) && !pte_protnone(pte)) {
pte               281 arch/powerpc/kvm/book3s_hv_rm_mmu.c 			if (writing && !__pte_write(pte))
pte               284 arch/powerpc/kvm/book3s_hv_rm_mmu.c 			is_ci = pte_ci(pte);
pte               285 arch/powerpc/kvm/book3s_hv_rm_mmu.c 			pa = pte_pfn(pte) << PAGE_SHIFT;
pte               338 arch/powerpc/kvm/book3s_hv_rm_mmu.c 				u64 pte;
pte               341 arch/powerpc/kvm/book3s_hv_rm_mmu.c 				pte = be64_to_cpu(hpte[0]);
pte               342 arch/powerpc/kvm/book3s_hv_rm_mmu.c 				if (!(pte & (HPTE_V_VALID | HPTE_V_ABSENT)))
pte               344 arch/powerpc/kvm/book3s_hv_rm_mmu.c 				__unlock_hpte(hpte, pte);
pte               356 arch/powerpc/kvm/book3s_hv_rm_mmu.c 			u64 pte;
pte               360 arch/powerpc/kvm/book3s_hv_rm_mmu.c 			pte = be64_to_cpu(hpte[0]);
pte               361 arch/powerpc/kvm/book3s_hv_rm_mmu.c 			if (pte & (HPTE_V_VALID | HPTE_V_ABSENT)) {
pte               362 arch/powerpc/kvm/book3s_hv_rm_mmu.c 				__unlock_hpte(hpte, pte);
pte               505 arch/powerpc/kvm/book3s_hv_rm_mmu.c 	u64 pte, orig_pte, pte_r;
pte               514 arch/powerpc/kvm/book3s_hv_rm_mmu.c 	pte = orig_pte = be64_to_cpu(hpte[0]);
pte               517 arch/powerpc/kvm/book3s_hv_rm_mmu.c 		pte = hpte_new_to_old_v(pte, pte_r);
pte               520 arch/powerpc/kvm/book3s_hv_rm_mmu.c 	if ((pte & (HPTE_V_ABSENT | HPTE_V_VALID)) == 0 ||
pte               521 arch/powerpc/kvm/book3s_hv_rm_mmu.c 	    ((flags & H_AVPN) && (pte & ~0x7fUL) != avpn) ||
pte               522 arch/powerpc/kvm/book3s_hv_rm_mmu.c 	    ((flags & H_ANDCOND) && (pte & avpn) != 0)) {
pte               528 arch/powerpc/kvm/book3s_hv_rm_mmu.c 	v = pte & ~HPTE_V_HVLOCK;
pte               899 arch/powerpc/kvm/book3s_hv_rm_mmu.c 	pte_t *ptep, pte;
pte               914 arch/powerpc/kvm/book3s_hv_rm_mmu.c 	pte = kvmppc_read_update_linux_pte(ptep, writing);
pte               915 arch/powerpc/kvm/book3s_hv_rm_mmu.c 	if (!pte_present(pte))
pte               921 arch/powerpc/kvm/book3s_hv_rm_mmu.c 	pa = pte_pfn(pte) << PAGE_SHIFT;
pte                56 arch/powerpc/kvm/book3s_mmu_hpte.c void kvmppc_mmu_hpte_cache_map(struct kvm_vcpu *vcpu, struct hpte_cache *pte)
pte                61 arch/powerpc/kvm/book3s_mmu_hpte.c 	trace_kvm_book3s_mmu_map(pte);
pte                66 arch/powerpc/kvm/book3s_mmu_hpte.c 	index = kvmppc_mmu_hash_pte(pte->pte.eaddr);
pte                67 arch/powerpc/kvm/book3s_mmu_hpte.c 	hlist_add_head_rcu(&pte->list_pte, &vcpu3s->hpte_hash_pte[index]);
pte                70 arch/powerpc/kvm/book3s_mmu_hpte.c 	index = kvmppc_mmu_hash_pte_long(pte->pte.eaddr);
pte                71 arch/powerpc/kvm/book3s_mmu_hpte.c 	hlist_add_head_rcu(&pte->list_pte_long,
pte                75 arch/powerpc/kvm/book3s_mmu_hpte.c 	index = kvmppc_mmu_hash_vpte(pte->pte.vpage);
pte                76 arch/powerpc/kvm/book3s_mmu_hpte.c 	hlist_add_head_rcu(&pte->list_vpte, &vcpu3s->hpte_hash_vpte[index]);
pte                79 arch/powerpc/kvm/book3s_mmu_hpte.c 	index = kvmppc_mmu_hash_vpte_long(pte->pte.vpage);
pte                80 arch/powerpc/kvm/book3s_mmu_hpte.c 	hlist_add_head_rcu(&pte->list_vpte_long,
pte                85 arch/powerpc/kvm/book3s_mmu_hpte.c 	index = kvmppc_mmu_hash_vpte_64k(pte->pte.vpage);
pte                86 arch/powerpc/kvm/book3s_mmu_hpte.c 	hlist_add_head_rcu(&pte->list_vpte_64k,
pte                97 arch/powerpc/kvm/book3s_mmu_hpte.c 	struct hpte_cache *pte = container_of(head, struct hpte_cache, rcu_head);
pte                98 arch/powerpc/kvm/book3s_mmu_hpte.c 	kmem_cache_free(hpte_cache, pte);
pte               101 arch/powerpc/kvm/book3s_mmu_hpte.c static void invalidate_pte(struct kvm_vcpu *vcpu, struct hpte_cache *pte)
pte               105 arch/powerpc/kvm/book3s_mmu_hpte.c 	trace_kvm_book3s_mmu_invalidate(pte);
pte               108 arch/powerpc/kvm/book3s_mmu_hpte.c 	kvmppc_mmu_invalidate_pte(vcpu, pte);
pte               113 arch/powerpc/kvm/book3s_mmu_hpte.c 	if (hlist_unhashed(&pte->list_pte)) {
pte               118 arch/powerpc/kvm/book3s_mmu_hpte.c 	hlist_del_init_rcu(&pte->list_pte);
pte               119 arch/powerpc/kvm/book3s_mmu_hpte.c 	hlist_del_init_rcu(&pte->list_pte_long);
pte               120 arch/powerpc/kvm/book3s_mmu_hpte.c 	hlist_del_init_rcu(&pte->list_vpte);
pte               121 arch/powerpc/kvm/book3s_mmu_hpte.c 	hlist_del_init_rcu(&pte->list_vpte_long);
pte               123 arch/powerpc/kvm/book3s_mmu_hpte.c 	hlist_del_init_rcu(&pte->list_vpte_64k);
pte               129 arch/powerpc/kvm/book3s_mmu_hpte.c 	call_rcu(&pte->rcu_head, free_pte_rcu);
pte               135 arch/powerpc/kvm/book3s_mmu_hpte.c 	struct hpte_cache *pte;
pte               143 arch/powerpc/kvm/book3s_mmu_hpte.c 		hlist_for_each_entry_rcu(pte, list, list_vpte_long)
pte               144 arch/powerpc/kvm/book3s_mmu_hpte.c 			invalidate_pte(vcpu, pte);
pte               154 arch/powerpc/kvm/book3s_mmu_hpte.c 	struct hpte_cache *pte;
pte               162 arch/powerpc/kvm/book3s_mmu_hpte.c 	hlist_for_each_entry_rcu(pte, list, list_pte)
pte               163 arch/powerpc/kvm/book3s_mmu_hpte.c 		if ((pte->pte.eaddr & ~0xfffUL) == guest_ea)
pte               164 arch/powerpc/kvm/book3s_mmu_hpte.c 			invalidate_pte(vcpu, pte);
pte               173 arch/powerpc/kvm/book3s_mmu_hpte.c 	struct hpte_cache *pte;
pte               182 arch/powerpc/kvm/book3s_mmu_hpte.c 	hlist_for_each_entry_rcu(pte, list, list_pte_long)
pte               183 arch/powerpc/kvm/book3s_mmu_hpte.c 		if ((pte->pte.eaddr & 0x0ffff000UL) == guest_ea)
pte               184 arch/powerpc/kvm/book3s_mmu_hpte.c 			invalidate_pte(vcpu, pte);
pte               216 arch/powerpc/kvm/book3s_mmu_hpte.c 	struct hpte_cache *pte;
pte               224 arch/powerpc/kvm/book3s_mmu_hpte.c 	hlist_for_each_entry_rcu(pte, list, list_vpte)
pte               225 arch/powerpc/kvm/book3s_mmu_hpte.c 		if ((pte->pte.vpage & vp_mask) == guest_vp)
pte               226 arch/powerpc/kvm/book3s_mmu_hpte.c 			invalidate_pte(vcpu, pte);
pte               237 arch/powerpc/kvm/book3s_mmu_hpte.c 	struct hpte_cache *pte;
pte               246 arch/powerpc/kvm/book3s_mmu_hpte.c 	hlist_for_each_entry_rcu(pte, list, list_vpte_64k)
pte               247 arch/powerpc/kvm/book3s_mmu_hpte.c 		if ((pte->pte.vpage & vp_mask) == guest_vp)
pte               248 arch/powerpc/kvm/book3s_mmu_hpte.c 			invalidate_pte(vcpu, pte);
pte               259 arch/powerpc/kvm/book3s_mmu_hpte.c 	struct hpte_cache *pte;
pte               268 arch/powerpc/kvm/book3s_mmu_hpte.c 	hlist_for_each_entry_rcu(pte, list, list_vpte_long)
pte               269 arch/powerpc/kvm/book3s_mmu_hpte.c 		if ((pte->pte.vpage & vp_mask) == guest_vp)
pte               270 arch/powerpc/kvm/book3s_mmu_hpte.c 			invalidate_pte(vcpu, pte);
pte               301 arch/powerpc/kvm/book3s_mmu_hpte.c 	struct hpte_cache *pte;
pte               311 arch/powerpc/kvm/book3s_mmu_hpte.c 		hlist_for_each_entry_rcu(pte, list, list_vpte_long)
pte               312 arch/powerpc/kvm/book3s_mmu_hpte.c 			if ((pte->pte.raddr >= pa_start) &&
pte               313 arch/powerpc/kvm/book3s_mmu_hpte.c 			    (pte->pte.raddr < pa_end))
pte               314 arch/powerpc/kvm/book3s_mmu_hpte.c 				invalidate_pte(vcpu, pte);
pte               323 arch/powerpc/kvm/book3s_mmu_hpte.c 	struct hpte_cache *pte;
pte               328 arch/powerpc/kvm/book3s_mmu_hpte.c 	pte = kmem_cache_zalloc(hpte_cache, GFP_KERNEL);
pte               330 arch/powerpc/kvm/book3s_mmu_hpte.c 	return pte;
pte               333 arch/powerpc/kvm/book3s_mmu_hpte.c void kvmppc_mmu_hpte_cache_free(struct hpte_cache *pte)
pte               335 arch/powerpc/kvm/book3s_mmu_hpte.c 	kmem_cache_free(hpte_cache, pte);
pte               443 arch/powerpc/kvm/book3s_pr.c static void kvm_set_spte_hva_pr(struct kvm *kvm, unsigned long hva, pte_t pte)
pte               625 arch/powerpc/kvm/book3s_pr.c static void kvmppc_patch_dcbz(struct kvm_vcpu *vcpu, struct kvmppc_pte *pte)
pte               632 arch/powerpc/kvm/book3s_pr.c 	hpage = gfn_to_page(vcpu->kvm, pte->raddr >> PAGE_SHIFT);
pte               636 arch/powerpc/kvm/book3s_pr.c 	hpage_offset = pte->raddr & ~PAGE_MASK;
pte               675 arch/powerpc/kvm/book3s_pr.c 	struct kvmppc_pte pte = { 0 };
pte               686 arch/powerpc/kvm/book3s_pr.c 		page_found = vcpu->arch.mmu.xlate(vcpu, eaddr, &pte, data, iswrite);
pte               688 arch/powerpc/kvm/book3s_pr.c 		pte.may_execute = true;
pte               689 arch/powerpc/kvm/book3s_pr.c 		pte.may_read = true;
pte               690 arch/powerpc/kvm/book3s_pr.c 		pte.may_write = true;
pte               691 arch/powerpc/kvm/book3s_pr.c 		pte.raddr = eaddr & KVM_PAM;
pte               692 arch/powerpc/kvm/book3s_pr.c 		pte.eaddr = eaddr;
pte               693 arch/powerpc/kvm/book3s_pr.c 		pte.vpage = eaddr >> 12;
pte               694 arch/powerpc/kvm/book3s_pr.c 		pte.page_size = MMU_PAGE_64K;
pte               695 arch/powerpc/kvm/book3s_pr.c 		pte.wimg = HPTE_R_M;
pte               700 arch/powerpc/kvm/book3s_pr.c 		pte.vpage |= ((u64)VSID_REAL << (SID_SHIFT - 12));
pte               705 arch/powerpc/kvm/book3s_pr.c 		    ((pte.raddr & SPLIT_HACK_MASK) == SPLIT_HACK_OFFS))
pte               706 arch/powerpc/kvm/book3s_pr.c 			pte.raddr &= ~SPLIT_HACK_MASK;
pte               712 arch/powerpc/kvm/book3s_pr.c 			pte.vpage |= ((u64)VSID_REAL_DR << (SID_SHIFT - 12));
pte               714 arch/powerpc/kvm/book3s_pr.c 			pte.vpage |= ((u64)VSID_REAL_IR << (SID_SHIFT - 12));
pte               715 arch/powerpc/kvm/book3s_pr.c 		pte.vpage |= vsid;
pte               729 arch/powerpc/kvm/book3s_pr.c 		pte.may_execute = !data;
pte               750 arch/powerpc/kvm/book3s_pr.c 	} else if (kvmppc_visible_gpa(vcpu, pte.raddr)) {
pte               757 arch/powerpc/kvm/book3s_pr.c 			kvmppc_mmu_unmap_page(vcpu, &pte);
pte               760 arch/powerpc/kvm/book3s_pr.c 		if (kvmppc_mmu_map_page(vcpu, &pte, iswrite) == -EIO) {
pte               769 arch/powerpc/kvm/book3s_pr.c 			kvmppc_patch_dcbz(vcpu, &pte);
pte               773 arch/powerpc/kvm/book3s_pr.c 		vcpu->arch.paddr_accessed = pte.raddr;
pte               774 arch/powerpc/kvm/book3s_pr.c 		vcpu->arch.vaddr_accessed = pte.eaddr;
pte                92 arch/powerpc/kvm/book3s_pr_papr.c 	unsigned long pte[2];
pte                98 arch/powerpc/kvm/book3s_pr_papr.c 	if (copy_from_user(pte, (void __user *)pteg, sizeof(pte)))
pte               100 arch/powerpc/kvm/book3s_pr_papr.c 	pte[0] = be64_to_cpu((__force __be64)pte[0]);
pte               101 arch/powerpc/kvm/book3s_pr_papr.c 	pte[1] = be64_to_cpu((__force __be64)pte[1]);
pte               104 arch/powerpc/kvm/book3s_pr_papr.c 	if ((pte[0] & HPTE_V_VALID) == 0 ||
pte               105 arch/powerpc/kvm/book3s_pr_papr.c 	    ((flags & H_AVPN) && (pte[0] & ~0x7fUL) != avpn) ||
pte               106 arch/powerpc/kvm/book3s_pr_papr.c 	    ((flags & H_ANDCOND) && (pte[0] & avpn) != 0))
pte               113 arch/powerpc/kvm/book3s_pr_papr.c 	rb = compute_tlbie_rb(pte[0], pte[1], pte_index);
pte               117 arch/powerpc/kvm/book3s_pr_papr.c 	kvmppc_set_gpr(vcpu, 4, pte[0]);
pte               118 arch/powerpc/kvm/book3s_pr_papr.c 	kvmppc_set_gpr(vcpu, 5, pte[1]);
pte               156 arch/powerpc/kvm/book3s_pr_papr.c 		unsigned long pte[2];
pte               179 arch/powerpc/kvm/book3s_pr_papr.c 		if (copy_from_user(pte, (void __user *)pteg, sizeof(pte))) {
pte               183 arch/powerpc/kvm/book3s_pr_papr.c 		pte[0] = be64_to_cpu((__force __be64)pte[0]);
pte               184 arch/powerpc/kvm/book3s_pr_papr.c 		pte[1] = be64_to_cpu((__force __be64)pte[1]);
pte               189 arch/powerpc/kvm/book3s_pr_papr.c 		if ((pte[0] & HPTE_V_VALID) == 0 ||
pte               190 arch/powerpc/kvm/book3s_pr_papr.c 		    ((flags & H_AVPN) && (pte[0] & ~0x7fUL) != tsl) ||
pte               191 arch/powerpc/kvm/book3s_pr_papr.c 		    ((flags & H_ANDCOND) && (pte[0] & tsl) != 0)) {
pte               200 arch/powerpc/kvm/book3s_pr_papr.c 			rb = compute_tlbie_rb(pte[0], pte[1],
pte               204 arch/powerpc/kvm/book3s_pr_papr.c 			tsh |= (pte[1] & (HPTE_R_C | HPTE_R_R)) << 43;
pte               220 arch/powerpc/kvm/book3s_pr_papr.c 	unsigned long pte[2];
pte               226 arch/powerpc/kvm/book3s_pr_papr.c 	if (copy_from_user(pte, (void __user *)pteg, sizeof(pte)))
pte               228 arch/powerpc/kvm/book3s_pr_papr.c 	pte[0] = be64_to_cpu((__force __be64)pte[0]);
pte               229 arch/powerpc/kvm/book3s_pr_papr.c 	pte[1] = be64_to_cpu((__force __be64)pte[1]);
pte               232 arch/powerpc/kvm/book3s_pr_papr.c 	if ((pte[0] & HPTE_V_VALID) == 0 ||
pte               233 arch/powerpc/kvm/book3s_pr_papr.c 	    ((flags & H_AVPN) && (pte[0] & ~0x7fUL) != avpn))
pte               236 arch/powerpc/kvm/book3s_pr_papr.c 	v = pte[0];
pte               237 arch/powerpc/kvm/book3s_pr_papr.c 	r = pte[1];
pte               244 arch/powerpc/kvm/book3s_pr_papr.c 	pte[1] = r;
pte               248 arch/powerpc/kvm/book3s_pr_papr.c 	pte[0] = (__force u64)cpu_to_be64(pte[0]);
pte               249 arch/powerpc/kvm/book3s_pr_papr.c 	pte[1] = (__force u64)cpu_to_be64(pte[1]);
pte               251 arch/powerpc/kvm/book3s_pr_papr.c 	if (copy_to_user((void __user *)pteg, pte, sizeof(pte)))
pte              1964 arch/powerpc/kvm/booke.c 		 enum xlate_readwrite xlrw, struct kvmppc_pte *pte)
pte              1972 arch/powerpc/kvm/booke.c 		pte->eaddr = eaddr;
pte              1973 arch/powerpc/kvm/booke.c 		pte->raddr = (vcpu->arch.magic_page_pa & PAGE_MASK) |
pte              1975 arch/powerpc/kvm/booke.c 		pte->vpage = eaddr >> PAGE_SHIFT;
pte              1976 arch/powerpc/kvm/booke.c 		pte->may_read = true;
pte              1977 arch/powerpc/kvm/booke.c 		pte->may_write = true;
pte              1978 arch/powerpc/kvm/booke.c 		pte->may_execute = true;
pte              2002 arch/powerpc/kvm/booke.c 	pte->eaddr = eaddr;
pte              2003 arch/powerpc/kvm/booke.c 	pte->raddr = (gpaddr & PAGE_MASK) | (eaddr & ~PAGE_MASK);
pte              2004 arch/powerpc/kvm/booke.c 	pte->vpage = eaddr >> PAGE_SHIFT;
pte              2007 arch/powerpc/kvm/booke.c 	pte->may_read = true;
pte              2008 arch/powerpc/kvm/booke.c 	pte->may_write = true;
pte              2009 arch/powerpc/kvm/booke.c 	pte->may_execute = true;
pte               479 arch/powerpc/kvm/e500_mmu_host.c 		pte_t pte = READ_ONCE(*ptep);
pte               481 arch/powerpc/kvm/e500_mmu_host.c 		if (pte_present(pte)) {
pte               482 arch/powerpc/kvm/e500_mmu_host.c 			wimg = (pte_val(pte) >> PTE_WIMGE_SHIFT) &
pte               757 arch/powerpc/kvm/e500_mmu_host.c int kvm_set_spte_hva(struct kvm *kvm, unsigned long hva, pte_t pte)
pte               327 arch/powerpc/kvm/powerpc.c 	struct kvmppc_pte pte;
pte               340 arch/powerpc/kvm/powerpc.c 			 XLATE_WRITE, &pte);
pte               344 arch/powerpc/kvm/powerpc.c 	*eaddr = pte.raddr;
pte               346 arch/powerpc/kvm/powerpc.c 	if (!pte.may_write)
pte               351 arch/powerpc/kvm/powerpc.c 	    ((pte.raddr & KVM_PAM & PAGE_MASK) == mp_pa) &&
pte               354 arch/powerpc/kvm/powerpc.c 		magic += pte.eaddr & 0xfff;
pte               359 arch/powerpc/kvm/powerpc.c 	if (kvm_write_guest(vcpu->kvm, pte.raddr, ptr, size))
pte               370 arch/powerpc/kvm/powerpc.c 	struct kvmppc_pte pte;
pte               383 arch/powerpc/kvm/powerpc.c 			  XLATE_READ, &pte);
pte               387 arch/powerpc/kvm/powerpc.c 	*eaddr = pte.raddr;
pte               389 arch/powerpc/kvm/powerpc.c 	if (!pte.may_read)
pte               392 arch/powerpc/kvm/powerpc.c 	if (!data && !pte.may_execute)
pte               397 arch/powerpc/kvm/powerpc.c 	    ((pte.raddr & KVM_PAM & PAGE_MASK) == mp_pa) &&
pte               400 arch/powerpc/kvm/powerpc.c 		magic += pte.eaddr & 0xfff;
pte               405 arch/powerpc/kvm/powerpc.c 	if (kvm_read_guest(vcpu->kvm, pte.raddr, ptr, size))
pte                64 arch/powerpc/kvm/trace_pr.h 	TP_PROTO(struct hpte_cache *pte),
pte                65 arch/powerpc/kvm/trace_pr.h 	TP_ARGS(pte),
pte                77 arch/powerpc/kvm/trace_pr.h 		__entry->host_vpn	= pte->host_vpn;
pte                78 arch/powerpc/kvm/trace_pr.h 		__entry->pfn		= pte->pfn;
pte                79 arch/powerpc/kvm/trace_pr.h 		__entry->eaddr		= pte->pte.eaddr;
pte                80 arch/powerpc/kvm/trace_pr.h 		__entry->vpage		= pte->pte.vpage;
pte                81 arch/powerpc/kvm/trace_pr.h 		__entry->raddr		= pte->pte.raddr;
pte                82 arch/powerpc/kvm/trace_pr.h 		__entry->flags		= (pte->pte.may_read ? 0x4 : 0) |
pte                83 arch/powerpc/kvm/trace_pr.h 					  (pte->pte.may_write ? 0x2 : 0) |
pte                84 arch/powerpc/kvm/trace_pr.h 					  (pte->pte.may_execute ? 0x1 : 0);
pte                93 arch/powerpc/kvm/trace_pr.h 	TP_PROTO(struct hpte_cache *pte),
pte                94 arch/powerpc/kvm/trace_pr.h 	TP_ARGS(pte),
pte               106 arch/powerpc/kvm/trace_pr.h 		__entry->host_vpn	= pte->host_vpn;
pte               107 arch/powerpc/kvm/trace_pr.h 		__entry->pfn		= pte->pfn;
pte               108 arch/powerpc/kvm/trace_pr.h 		__entry->eaddr		= pte->pte.eaddr;
pte               109 arch/powerpc/kvm/trace_pr.h 		__entry->vpage		= pte->pte.vpage;
pte               110 arch/powerpc/kvm/trace_pr.h 		__entry->raddr		= pte->pte.raddr;
pte               111 arch/powerpc/kvm/trace_pr.h 		__entry->flags		= (pte->pte.may_read ? 0x4 : 0) |
pte               112 arch/powerpc/kvm/trace_pr.h 					  (pte->pte.may_write ? 0x2 : 0) |
pte               113 arch/powerpc/kvm/trace_pr.h 					  (pte->pte.may_execute ? 0x1 : 0);
pte                34 arch/powerpc/mm/book3s64/hash_4k.c 		pte_t pte = READ_ONCE(*ptep);
pte                36 arch/powerpc/mm/book3s64/hash_4k.c 		old_pte = pte_val(pte);
pte                52 arch/powerpc/mm/book3s64/hash_64k.c 		pte_t pte = READ_ONCE(*ptep);
pte                54 arch/powerpc/mm/book3s64/hash_64k.c 		old_pte = pte_val(pte);
pte               238 arch/powerpc/mm/book3s64/hash_64k.c 		pte_t pte = READ_ONCE(*ptep);
pte               240 arch/powerpc/mm/book3s64/hash_64k.c 		old_pte = pte_val(pte);
pte               252 arch/powerpc/mm/book3s64/hash_64k.c 		    unlikely(pte_ci(pte)))
pte               148 arch/powerpc/mm/book3s64/hash_hugetlbpage.c 				  pte_t *ptep, pte_t old_pte, pte_t pte)
pte               153 arch/powerpc/mm/book3s64/hash_hugetlbpage.c 							   old_pte, pte);
pte               154 arch/powerpc/mm/book3s64/hash_hugetlbpage.c 	set_huge_pte_at(vma->vm_mm, addr, ptep, pte);
pte               810 arch/powerpc/mm/book3s64/hash_native.c 	real_pte_t pte;
pte               824 arch/powerpc/mm/book3s64/hash_native.c 		pte = batch->pte[i];
pte               826 arch/powerpc/mm/book3s64/hash_native.c 		pte_iterate_hashed_subpages(pte, psize, vpn, index, shift) {
pte               828 arch/powerpc/mm/book3s64/hash_native.c 			hidx = __rpte_to_hidx(pte, index);
pte               855 arch/powerpc/mm/book3s64/hash_native.c 			pte = batch->pte[i];
pte               857 arch/powerpc/mm/book3s64/hash_native.c 			pte_iterate_hashed_subpages(pte, psize,
pte               872 arch/powerpc/mm/book3s64/hash_native.c 			pte = batch->pte[i];
pte               874 arch/powerpc/mm/book3s64/hash_native.c 			pte_iterate_hashed_subpages(pte, psize,
pte                42 arch/powerpc/mm/book3s64/hash_tlb.c 		     pte_t *ptep, unsigned long pte, int huge)
pte                73 arch/powerpc/mm/book3s64/hash_tlb.c 		psize = pte_pagesize_index(mm, addr, pte); /* shutup gcc */
pte                76 arch/powerpc/mm/book3s64/hash_tlb.c 		psize = pte_pagesize_index(mm, addr, pte);
pte                98 arch/powerpc/mm/book3s64/hash_tlb.c 	rpte = __real_pte(__pte(pte), ptep, offset);
pte               130 arch/powerpc/mm/book3s64/hash_tlb.c 	batch->pte[i] = rpte;
pte               152 arch/powerpc/mm/book3s64/hash_tlb.c 		flush_hash_page(batch->vpn[0], batch->pte[0],
pte               217 arch/powerpc/mm/book3s64/hash_tlb.c 		unsigned long pte;
pte               221 arch/powerpc/mm/book3s64/hash_tlb.c 		pte = pte_val(*ptep);
pte               223 arch/powerpc/mm/book3s64/hash_tlb.c 			trace_hugepage_invalidate(start, pte);
pte               224 arch/powerpc/mm/book3s64/hash_tlb.c 		if (!(pte & H_PAGE_HASHPTE))
pte               227 arch/powerpc/mm/book3s64/hash_tlb.c 			hpte_do_hugepage_flush(mm, start, (pmd_t *)ptep, pte);
pte               229 arch/powerpc/mm/book3s64/hash_tlb.c 			hpte_need_flush(mm, start, ptep, pte, hugepage_shift);
pte               237 arch/powerpc/mm/book3s64/hash_tlb.c 	pte_t *pte;
pte               253 arch/powerpc/mm/book3s64/hash_tlb.c 	for (pte = start_pte; pte < start_pte + PTRS_PER_PTE; pte++) {
pte               254 arch/powerpc/mm/book3s64/hash_tlb.c 		unsigned long pteval = pte_val(*pte);
pte               256 arch/powerpc/mm/book3s64/hash_tlb.c 			hpte_need_flush(mm, addr, pte, pteval, 0);
pte              1093 arch/powerpc/mm/book3s64/hash_utils.c unsigned int hash_page_do_lazy_icache(unsigned int pp, pte_t pte, int trap)
pte              1097 arch/powerpc/mm/book3s64/hash_utils.c 	if (!pfn_valid(pte_pfn(pte)))
pte              1100 arch/powerpc/mm/book3s64/hash_utils.c 	page = pte_page(pte);
pte              1210 arch/powerpc/mm/book3s64/hash_utils.c 			int ssize, int psize, int lpsize, unsigned long pte)
pte              1217 arch/powerpc/mm/book3s64/hash_utils.c 		trap, vsid, ssize, psize, lpsize, pte);
pte              1722 arch/powerpc/mm/book3s64/hash_utils.c void flush_hash_page(unsigned long vpn, real_pte_t pte, int psize, int ssize,
pte              1729 arch/powerpc/mm/book3s64/hash_utils.c 	pte_iterate_hashed_subpages(pte, psize, vpn, index, shift) {
pte              1730 arch/powerpc/mm/book3s64/hash_utils.c 		gslot = pte_get_hash_gslot(vpn, shift, ssize, pte, index);
pte              1811 arch/powerpc/mm/book3s64/hash_utils.c 			flush_hash_page(batch->vpn[i], batch->pte[i],
pte               437 arch/powerpc/mm/book3s64/pgtable.c 			     pte_t *ptep, pte_t old_pte, pte_t pte)
pte               441 arch/powerpc/mm/book3s64/pgtable.c 						      ptep, old_pte, pte);
pte               442 arch/powerpc/mm/book3s64/pgtable.c 	set_pte_at(vma->vm_mm, addr, ptep, pte);
pte               378 arch/powerpc/mm/book3s64/pkeys.c bool arch_pte_access_permitted(u64 pte, bool write, bool execute)
pte               383 arch/powerpc/mm/book3s64/pkeys.c 	return pkey_access_permitted(pte_to_pkey_bits(pte), write, execute);
pte                97 arch/powerpc/mm/book3s64/radix_hugetlbpage.c 					 pte_t old_pte, pte_t pte)
pte               105 arch/powerpc/mm/book3s64/radix_hugetlbpage.c 	if (is_pte_rw_upgrade(pte_val(old_pte), pte_val(pte)) &&
pte               109 arch/powerpc/mm/book3s64/radix_hugetlbpage.c 	set_huge_pte_at(vma->vm_mm, addr, ptep, pte);
pte               664 arch/powerpc/mm/book3s64/radix_pgtable.c 	pte_t *pte;
pte               668 arch/powerpc/mm/book3s64/radix_pgtable.c 		pte = pte_start + i;
pte               669 arch/powerpc/mm/book3s64/radix_pgtable.c 		if (!pte_none(*pte))
pte               693 arch/powerpc/mm/book3s64/radix_pgtable.c 	pte_t *pte;
pte               709 arch/powerpc/mm/book3s64/radix_pgtable.c 	pte_clear(&init_mm, params->aligned_start, params->pte);
pte               720 arch/powerpc/mm/book3s64/radix_pgtable.c 	pte_t *pte;
pte               722 arch/powerpc/mm/book3s64/radix_pgtable.c 	pte = pte_start + pte_index(addr);
pte               723 arch/powerpc/mm/book3s64/radix_pgtable.c 	for (; addr < end; addr = next, pte++) {
pte               728 arch/powerpc/mm/book3s64/radix_pgtable.c 		if (!pte_present(*pte))
pte               740 arch/powerpc/mm/book3s64/radix_pgtable.c 		pte_clear(&init_mm, addr, pte);
pte               748 arch/powerpc/mm/book3s64/radix_pgtable.c 				unsigned long size, pte_t *pte)
pte               777 arch/powerpc/mm/book3s64/radix_pgtable.c 		params.pte = pte;
pte               787 arch/powerpc/mm/book3s64/radix_pgtable.c 	pte_clear(&init_mm, addr, pte);
pte              1068 arch/powerpc/mm/book3s64/radix_pgtable.c 				    pte_t old_pte, pte_t pte)
pte              1077 arch/powerpc/mm/book3s64/radix_pgtable.c 	if (is_pte_rw_upgrade(pte_val(old_pte), pte_val(pte)) &&
pte              1081 arch/powerpc/mm/book3s64/radix_pgtable.c 	set_pte_at(mm, addr, ptep, pte);
pte              1135 arch/powerpc/mm/book3s64/radix_pgtable.c 			pte_t *pte;
pte              1136 arch/powerpc/mm/book3s64/radix_pgtable.c 			pte = (pte_t *)pmd_page_vaddr(pmd[i]);
pte              1138 arch/powerpc/mm/book3s64/radix_pgtable.c 			pte_free_kernel(&init_mm, pte);
pte              1172 arch/powerpc/mm/book3s64/radix_pgtable.c 	pte_t *pte;
pte              1174 arch/powerpc/mm/book3s64/radix_pgtable.c 	pte = (pte_t *)pmd_page_vaddr(*pmd);
pte              1179 arch/powerpc/mm/book3s64/radix_pgtable.c 	pte_free_kernel(&init_mm, pte);
pte                59 arch/powerpc/mm/book3s64/subpage_prot.c 	pte_t *pte;
pte                71 arch/powerpc/mm/book3s64/subpage_prot.c 	pte = pte_offset_map_lock(mm, pmd, addr, &ptl);
pte                74 arch/powerpc/mm/book3s64/subpage_prot.c 		pte_update(mm, addr, pte, 0, 0, 0);
pte                76 arch/powerpc/mm/book3s64/subpage_prot.c 		++pte;
pte                79 arch/powerpc/mm/book3s64/subpage_prot.c 	pte_unmap_unlock(pte - 1, ptl);
pte                45 arch/powerpc/mm/ioremap.c 	pte_t pte = __pte(flags);
pte                49 arch/powerpc/mm/ioremap.c 	if (pte_write(pte))
pte                50 arch/powerpc/mm/ioremap.c 		pte = pte_mkdirty(pte);
pte                53 arch/powerpc/mm/ioremap.c 	pte = pte_exprotect(pte);
pte                54 arch/powerpc/mm/ioremap.c 	pte = pte_mkprivileged(pte);
pte                57 arch/powerpc/mm/ioremap.c 		return iowa_ioremap(addr, size, pte_pgprot(pte), caller);
pte                58 arch/powerpc/mm/ioremap.c 	return __ioremap_caller(addr, size, pte_pgprot(pte), caller);
pte                99 arch/powerpc/mm/kasan/kasan_init_32.c 		pte_t pte = pfn_pte(PHYS_PFN(__pa(va)), PAGE_KERNEL);
pte               104 arch/powerpc/mm/kasan/kasan_init_32.c 		__set_pte_at(&init_mm, k_cur, pte_offset_kernel(pmd, k_cur), pte, 0);
pte               126 arch/powerpc/mm/nohash/book3e_hugetlbpage.c book3e_hugetlb_preload(struct vm_area_struct *vma, unsigned long ea, pte_t pte)
pte               163 arch/powerpc/mm/nohash/book3e_hugetlbpage.c 	mas2 |= (pte_val(pte) >> PTE_WIMGE_SHIFT) & MAS2_WIMGE_MASK;
pte               164 arch/powerpc/mm/nohash/book3e_hugetlbpage.c 	mas7_3 = (u64)pte_pfn(pte) << PAGE_SHIFT;
pte               165 arch/powerpc/mm/nohash/book3e_hugetlbpage.c 	mas7_3 |= (pte_val(pte) >> PTE_BAP_SHIFT) & MAS3_BAP_MASK;
pte               166 arch/powerpc/mm/nohash/book3e_hugetlbpage.c 	if (!pte_dirty(pte))
pte               100 arch/powerpc/mm/pgtable-frag.c 	pte_t *pte;
pte               102 arch/powerpc/mm/pgtable-frag.c 	pte = get_pte_from_cache(mm);
pte               103 arch/powerpc/mm/pgtable-frag.c 	if (pte)
pte               104 arch/powerpc/mm/pgtable-frag.c 		return pte;
pte                41 arch/powerpc/mm/pgtable.c static inline int pte_looks_normal(pte_t pte)
pte                44 arch/powerpc/mm/pgtable.c 	if (pte_present(pte) && !pte_special(pte)) {
pte                45 arch/powerpc/mm/pgtable.c 		if (pte_ci(pte))
pte                47 arch/powerpc/mm/pgtable.c 		if (pte_user(pte))
pte                53 arch/powerpc/mm/pgtable.c static struct page *maybe_pte_to_page(pte_t pte)
pte                55 arch/powerpc/mm/pgtable.c 	unsigned long pfn = pte_pfn(pte);
pte                74 arch/powerpc/mm/pgtable.c static pte_t set_pte_filter_hash(pte_t pte)
pte                77 arch/powerpc/mm/pgtable.c 		return pte;
pte                79 arch/powerpc/mm/pgtable.c 	pte = __pte(pte_val(pte) & ~_PAGE_HPTEFLAGS);
pte                80 arch/powerpc/mm/pgtable.c 	if (pte_looks_normal(pte) && !(cpu_has_feature(CPU_FTR_COHERENT_ICACHE) ||
pte                82 arch/powerpc/mm/pgtable.c 		struct page *pg = maybe_pte_to_page(pte);
pte                84 arch/powerpc/mm/pgtable.c 			return pte;
pte                90 arch/powerpc/mm/pgtable.c 	return pte;
pte                95 arch/powerpc/mm/pgtable.c static pte_t set_pte_filter_hash(pte_t pte) { return pte; }
pte               103 arch/powerpc/mm/pgtable.c static pte_t set_pte_filter(pte_t pte)
pte               108 arch/powerpc/mm/pgtable.c 		return set_pte_filter_hash(pte);
pte               111 arch/powerpc/mm/pgtable.c 	if (!pte_exec(pte) || !pte_looks_normal(pte))
pte               112 arch/powerpc/mm/pgtable.c 		return pte;
pte               115 arch/powerpc/mm/pgtable.c 	pg = maybe_pte_to_page(pte);
pte               117 arch/powerpc/mm/pgtable.c 		return pte;
pte               121 arch/powerpc/mm/pgtable.c 		return pte;
pte               127 arch/powerpc/mm/pgtable.c 		return pte;
pte               131 arch/powerpc/mm/pgtable.c 	return pte_exprotect(pte);
pte               134 arch/powerpc/mm/pgtable.c static pte_t set_access_flags_filter(pte_t pte, struct vm_area_struct *vma,
pte               140 arch/powerpc/mm/pgtable.c 		return pte;
pte               147 arch/powerpc/mm/pgtable.c 	if (dirty || pte_exec(pte) || !is_exec_fault())
pte               148 arch/powerpc/mm/pgtable.c 		return pte;
pte               156 arch/powerpc/mm/pgtable.c 		return pte;
pte               160 arch/powerpc/mm/pgtable.c 	pg = maybe_pte_to_page(pte);
pte               173 arch/powerpc/mm/pgtable.c 	return pte_mkexec(pte);
pte               180 arch/powerpc/mm/pgtable.c 		pte_t pte)
pte               189 arch/powerpc/mm/pgtable.c 	pte = pte_mkpte(pte);
pte               195 arch/powerpc/mm/pgtable.c 	pte = set_pte_filter(pte);
pte               198 arch/powerpc/mm/pgtable.c 	__set_pte_at(mm, addr, ptep, pte, 0);
pte               225 arch/powerpc/mm/pgtable.c 			       pte_t pte, int dirty)
pte               233 arch/powerpc/mm/pgtable.c 	ptep_set_access_flags(vma, addr, ptep, pte, dirty);
pte               238 arch/powerpc/mm/pgtable.c 	pte = set_access_flags_filter(pte, vma, dirty);
pte               239 arch/powerpc/mm/pgtable.c 	changed = !pte_same(*(ptep), pte);
pte               257 arch/powerpc/mm/pgtable.c 		__ptep_set_access_flags(vma, ptep, pte, addr, psize);
pte               138 arch/powerpc/mm/pgtable_32.c         pte_t	*pte;
pte               147 arch/powerpc/mm/pgtable_32.c 				pte = pte_offset_map(pmd, addr & PAGE_MASK);
pte               148 arch/powerpc/mm/pgtable_32.c 				if (pte) {
pte               150 arch/powerpc/mm/pgtable_32.c 					*ptep = pte;
pte               156 arch/powerpc/mm/ptdump/hashpagetable.c 		*flag, u64 pte, int num)
pte               169 arch/powerpc/mm/ptdump/hashpagetable.c 			val = pte & flag->val;
pte               174 arch/powerpc/mm/ptdump/hashpagetable.c 			if ((pte & flag->mask) == flag->val)
pte               374 arch/powerpc/mm/ptdump/hashpagetable.c 	pte_t *pte = pte_offset_kernel(pmd, 0);
pte               378 arch/powerpc/mm/ptdump/hashpagetable.c 	for (i = 0; i < PTRS_PER_PTE; i++, pte++) {
pte               380 arch/powerpc/mm/ptdump/hashpagetable.c 		pteval = pte_val(*pte);
pte               115 arch/powerpc/mm/ptdump/ptdump.c 		*flag, u64 pte, int num)
pte               128 arch/powerpc/mm/ptdump/ptdump.c 			val = pte & flag->val;
pte               133 arch/powerpc/mm/ptdump/ptdump.c 			if ((pte & flag->mask) == flag->val)
pte               177 arch/powerpc/mm/ptdump/ptdump.c 	pte_t pte = __pte(st->current_flags);
pte               182 arch/powerpc/mm/ptdump/ptdump.c 	if (!pte_write(pte) || !pte_exec(pte))
pte               254 arch/powerpc/mm/ptdump/ptdump.c 	pte_t *pte = pte_offset_kernel(pmd, 0);
pte               258 arch/powerpc/mm/ptdump/ptdump.c 	for (i = 0; i < PTRS_PER_PTE; i++, pte++) {
pte               260 arch/powerpc/mm/ptdump/ptdump.c 		note_page(st, addr, 4, pte_val(*pte), PAGE_SIZE);
pte               115 arch/powerpc/perf/callchain.c 	pte_t *ptep, pte;
pte               136 arch/powerpc/perf/callchain.c 	pte = READ_ONCE(*ptep);
pte               137 arch/powerpc/perf/callchain.c 	if (!pte_present(pte) || !pte_user(pte))
pte               139 arch/powerpc/perf/callchain.c 	pfn = pte_pfn(pte);
pte               129 arch/powerpc/platforms/cell/iommu.c static void invalidate_tce_cache(struct cbe_iommu *iommu, unsigned long *pte,
pte               142 arch/powerpc/platforms/cell/iommu.c 			| (__pa(pte) & IOC_IOPT_CacheInvd_IOPTE_Mask)
pte               150 arch/powerpc/platforms/cell/iommu.c 		pte += n;
pte               204 arch/powerpc/platforms/cell/iommu.c 	unsigned long *io_pte, pte;
pte               211 arch/powerpc/platforms/cell/iommu.c 	pte = 0;
pte               215 arch/powerpc/platforms/cell/iommu.c 	pte = CBE_IOPTE_PP_R | CBE_IOPTE_M | CBE_IOPTE_SO_RW |
pte               223 arch/powerpc/platforms/cell/iommu.c 		io_pte[i] = pte;
pte              1263 arch/powerpc/platforms/pseries/lpar.c static inline unsigned long compute_slot(real_pte_t pte,
pte              1272 arch/powerpc/platforms/pseries/lpar.c 	hidx = __rpte_to_hidx(pte, index);
pte              1290 arch/powerpc/platforms/pseries/lpar.c 	real_pte_t pte;
pte              1298 arch/powerpc/platforms/pseries/lpar.c 		pte = batch->pte[i];
pte              1299 arch/powerpc/platforms/pseries/lpar.c 		pte_iterate_hashed_subpages(pte, psize, vpn, index, shift) {
pte              1319 arch/powerpc/platforms/pseries/lpar.c 			slot = compute_slot(pte, vpn, index, shift, ssize);
pte              1489 arch/powerpc/platforms/pseries/lpar.c 	real_pte_t pte;
pte              1505 arch/powerpc/platforms/pseries/lpar.c 		pte = batch->pte[i];
pte              1506 arch/powerpc/platforms/pseries/lpar.c 		pte_iterate_hashed_subpages(pte, psize, vpn, index, shift) {
pte              1507 arch/powerpc/platforms/pseries/lpar.c 			slot = compute_slot(pte, vpn, index, shift, ssize);
pte              3103 arch/powerpc/xmon/xmon.c static void format_pte(void *ptep, unsigned long pte)
pte              3105 arch/powerpc/xmon/xmon.c 	pte_t entry = __pte(pte);
pte              3107 arch/powerpc/xmon/xmon.c 	printf("ptep @ 0x%016lx = 0x%016lx\n", (unsigned long)ptep, pte);
pte              3108 arch/powerpc/xmon/xmon.c 	printf("Maps physical address = 0x%016lx\n", pte & PTE_RPN_MASK);
pte                68 arch/riscv/include/asm/page.h 	unsigned long pte;
pte                77 arch/riscv/include/asm/page.h #define pte_val(x)	((x).pte)
pte                16 arch/riscv/include/asm/pgalloc.h 	pmd_t *pmd, pte_t *pte)
pte                18 arch/riscv/include/asm/pgalloc.h 	unsigned long pfn = virt_to_pfn(pte);
pte                24 arch/riscv/include/asm/pgalloc.h 	pmd_t *pmd, pgtable_t pte)
pte                26 arch/riscv/include/asm/pgalloc.h 	unsigned long pfn = virt_to_pfn(page_address(pte));
pte                79 arch/riscv/include/asm/pgalloc.h #define __pte_free_tlb(tlb, pte, buf)   \
pte                81 arch/riscv/include/asm/pgalloc.h 	pgtable_pte_page_dtor(pte);     \
pte                82 arch/riscv/include/asm/pgalloc.h 	tlb_remove_page((tlb), pte);    \
pte               179 arch/riscv/include/asm/pgtable.h static inline unsigned long pte_pfn(pte_t pte)
pte               181 arch/riscv/include/asm/pgtable.h 	return (pte_val(pte) >> _PAGE_PFN_SHIFT);
pte               202 arch/riscv/include/asm/pgtable.h #define pte_unmap(pte)			((void)(pte))
pte               204 arch/riscv/include/asm/pgtable.h static inline int pte_present(pte_t pte)
pte               206 arch/riscv/include/asm/pgtable.h 	return (pte_val(pte) & (_PAGE_PRESENT | _PAGE_PROT_NONE));
pte               209 arch/riscv/include/asm/pgtable.h static inline int pte_none(pte_t pte)
pte               211 arch/riscv/include/asm/pgtable.h 	return (pte_val(pte) == 0);
pte               214 arch/riscv/include/asm/pgtable.h static inline int pte_write(pte_t pte)
pte               216 arch/riscv/include/asm/pgtable.h 	return pte_val(pte) & _PAGE_WRITE;
pte               219 arch/riscv/include/asm/pgtable.h static inline int pte_exec(pte_t pte)
pte               221 arch/riscv/include/asm/pgtable.h 	return pte_val(pte) & _PAGE_EXEC;
pte               224 arch/riscv/include/asm/pgtable.h static inline int pte_huge(pte_t pte)
pte               226 arch/riscv/include/asm/pgtable.h 	return pte_present(pte)
pte               227 arch/riscv/include/asm/pgtable.h 		&& (pte_val(pte) & (_PAGE_READ | _PAGE_WRITE | _PAGE_EXEC));
pte               230 arch/riscv/include/asm/pgtable.h static inline int pte_dirty(pte_t pte)
pte               232 arch/riscv/include/asm/pgtable.h 	return pte_val(pte) & _PAGE_DIRTY;
pte               235 arch/riscv/include/asm/pgtable.h static inline int pte_young(pte_t pte)
pte               237 arch/riscv/include/asm/pgtable.h 	return pte_val(pte) & _PAGE_ACCESSED;
pte               240 arch/riscv/include/asm/pgtable.h static inline int pte_special(pte_t pte)
pte               242 arch/riscv/include/asm/pgtable.h 	return pte_val(pte) & _PAGE_SPECIAL;
pte               247 arch/riscv/include/asm/pgtable.h static inline pte_t pte_wrprotect(pte_t pte)
pte               249 arch/riscv/include/asm/pgtable.h 	return __pte(pte_val(pte) & ~(_PAGE_WRITE));
pte               254 arch/riscv/include/asm/pgtable.h static inline pte_t pte_mkwrite(pte_t pte)
pte               256 arch/riscv/include/asm/pgtable.h 	return __pte(pte_val(pte) | _PAGE_WRITE);
pte               261 arch/riscv/include/asm/pgtable.h static inline pte_t pte_mkdirty(pte_t pte)
pte               263 arch/riscv/include/asm/pgtable.h 	return __pte(pte_val(pte) | _PAGE_DIRTY);
pte               266 arch/riscv/include/asm/pgtable.h static inline pte_t pte_mkclean(pte_t pte)
pte               268 arch/riscv/include/asm/pgtable.h 	return __pte(pte_val(pte) & ~(_PAGE_DIRTY));
pte               271 arch/riscv/include/asm/pgtable.h static inline pte_t pte_mkyoung(pte_t pte)
pte               273 arch/riscv/include/asm/pgtable.h 	return __pte(pte_val(pte) | _PAGE_ACCESSED);
pte               276 arch/riscv/include/asm/pgtable.h static inline pte_t pte_mkold(pte_t pte)
pte               278 arch/riscv/include/asm/pgtable.h 	return __pte(pte_val(pte) & ~(_PAGE_ACCESSED));
pte               281 arch/riscv/include/asm/pgtable.h static inline pte_t pte_mkspecial(pte_t pte)
pte               283 arch/riscv/include/asm/pgtable.h 	return __pte(pte_val(pte) | _PAGE_SPECIAL);
pte               286 arch/riscv/include/asm/pgtable.h static inline pte_t pte_mkhuge(pte_t pte)
pte               288 arch/riscv/include/asm/pgtable.h 	return pte;
pte               292 arch/riscv/include/asm/pgtable.h static inline pte_t pte_modify(pte_t pte, pgprot_t newprot)
pte               294 arch/riscv/include/asm/pgtable.h 	return __pte((pte_val(pte) & _PAGE_CHG_MASK) | pgprot_val(newprot));
pte               331 arch/riscv/include/asm/pgtable.h void flush_icache_pte(pte_t pte);
pte               430 arch/riscv/include/asm/pgtable.h #define __pte_to_swp_entry(pte)	((swp_entry_t) { pte_val(pte) })
pte                70 arch/riscv/mm/cacheflush.c void flush_icache_pte(pte_t pte)
pte                72 arch/riscv/mm/cacheflush.c 	struct page *page = pte_page(pte);
pte               135 arch/s390/include/asm/gmap.h int gmap_shadow_page(struct gmap *sg, unsigned long saddr, pte_t pte);
pte                19 arch/s390/include/asm/hugetlb.h 		     pte_t *ptep, pte_t pte);
pte                67 arch/s390/include/asm/hugetlb.h 					     pte_t pte, int dirty)
pte                69 arch/s390/include/asm/hugetlb.h 	int changed = !pte_same(huge_ptep_get(ptep), pte);
pte                72 arch/s390/include/asm/hugetlb.h 		set_huge_pte_at(vma->vm_mm, addr, ptep, pte);
pte                80 arch/s390/include/asm/hugetlb.h 	pte_t pte = huge_ptep_get_and_clear(mm, addr, ptep);
pte                81 arch/s390/include/asm/hugetlb.h 	set_huge_pte_at(mm, addr, ptep, pte_wrprotect(pte));
pte                89 arch/s390/include/asm/hugetlb.h static inline int huge_pte_none(pte_t pte)
pte                91 arch/s390/include/asm/hugetlb.h 	return pte_none(pte);
pte                94 arch/s390/include/asm/hugetlb.h static inline int huge_pte_write(pte_t pte)
pte                96 arch/s390/include/asm/hugetlb.h 	return pte_write(pte);
pte                99 arch/s390/include/asm/hugetlb.h static inline int huge_pte_dirty(pte_t pte)
pte               101 arch/s390/include/asm/hugetlb.h 	return pte_dirty(pte);
pte               104 arch/s390/include/asm/hugetlb.h static inline pte_t huge_pte_mkwrite(pte_t pte)
pte               106 arch/s390/include/asm/hugetlb.h 	return pte_mkwrite(pte);
pte               109 arch/s390/include/asm/hugetlb.h static inline pte_t huge_pte_mkdirty(pte_t pte)
pte               111 arch/s390/include/asm/hugetlb.h 	return pte_mkdirty(pte);
pte               114 arch/s390/include/asm/hugetlb.h static inline pte_t huge_pte_wrprotect(pte_t pte)
pte               116 arch/s390/include/asm/hugetlb.h 	return pte_wrprotect(pte);
pte               119 arch/s390/include/asm/hugetlb.h static inline pte_t huge_pte_modify(pte_t pte, pgprot_t newprot)
pte               121 arch/s390/include/asm/hugetlb.h 	return pte_modify(pte, newprot);
pte                81 arch/s390/include/asm/page.h typedef struct { unsigned long pte; } pte_t;
pte                90 arch/s390/include/asm/page.h #define pte_val(x)	((x).pte)
pte               141 arch/s390/include/asm/pgalloc.h 				pmd_t *pmd, pgtable_t pte)
pte               143 arch/s390/include/asm/pgalloc.h 	pmd_val(*pmd) = _SEGMENT_ENTRY + __pa(pte);
pte               146 arch/s390/include/asm/pgalloc.h #define pmd_populate_kernel(mm, pmd, pte) pmd_populate(mm, pmd, pte)
pte               157 arch/s390/include/asm/pgalloc.h #define pte_free_kernel(mm, pte) page_table_free(mm, (unsigned long *) pte)
pte               158 arch/s390/include/asm/pgalloc.h #define pte_free(mm, pte) page_table_free(mm, (unsigned long *) pte)
pte               781 arch/s390/include/asm/pgtable.h static inline int pte_present(pte_t pte)
pte               784 arch/s390/include/asm/pgtable.h 	return (pte_val(pte) & _PAGE_PRESENT) != 0;
pte               787 arch/s390/include/asm/pgtable.h static inline int pte_none(pte_t pte)
pte               790 arch/s390/include/asm/pgtable.h 	return pte_val(pte) == _PAGE_INVALID;
pte               793 arch/s390/include/asm/pgtable.h static inline int pte_swap(pte_t pte)
pte               796 arch/s390/include/asm/pgtable.h 	return (pte_val(pte) & (_PAGE_PROTECT | _PAGE_PRESENT))
pte               800 arch/s390/include/asm/pgtable.h static inline int pte_special(pte_t pte)
pte               802 arch/s390/include/asm/pgtable.h 	return (pte_val(pte) & _PAGE_SPECIAL);
pte               812 arch/s390/include/asm/pgtable.h static inline int pte_protnone(pte_t pte)
pte               814 arch/s390/include/asm/pgtable.h 	return pte_present(pte) && !(pte_val(pte) & _PAGE_READ);
pte               824 arch/s390/include/asm/pgtable.h static inline int pte_soft_dirty(pte_t pte)
pte               826 arch/s390/include/asm/pgtable.h 	return pte_val(pte) & _PAGE_SOFT_DIRTY;
pte               830 arch/s390/include/asm/pgtable.h static inline pte_t pte_mksoft_dirty(pte_t pte)
pte               832 arch/s390/include/asm/pgtable.h 	pte_val(pte) |= _PAGE_SOFT_DIRTY;
pte               833 arch/s390/include/asm/pgtable.h 	return pte;
pte               837 arch/s390/include/asm/pgtable.h static inline pte_t pte_clear_soft_dirty(pte_t pte)
pte               839 arch/s390/include/asm/pgtable.h 	pte_val(pte) &= ~_PAGE_SOFT_DIRTY;
pte               840 arch/s390/include/asm/pgtable.h 	return pte;
pte               865 arch/s390/include/asm/pgtable.h static inline int pte_write(pte_t pte)
pte               867 arch/s390/include/asm/pgtable.h 	return (pte_val(pte) & _PAGE_WRITE) != 0;
pte               870 arch/s390/include/asm/pgtable.h static inline int pte_dirty(pte_t pte)
pte               872 arch/s390/include/asm/pgtable.h 	return (pte_val(pte) & _PAGE_DIRTY) != 0;
pte               875 arch/s390/include/asm/pgtable.h static inline int pte_young(pte_t pte)
pte               877 arch/s390/include/asm/pgtable.h 	return (pte_val(pte) & _PAGE_YOUNG) != 0;
pte               881 arch/s390/include/asm/pgtable.h static inline int pte_unused(pte_t pte)
pte               883 arch/s390/include/asm/pgtable.h 	return pte_val(pte) & _PAGE_UNUSED;
pte               922 arch/s390/include/asm/pgtable.h static inline pte_t pte_modify(pte_t pte, pgprot_t newprot)
pte               924 arch/s390/include/asm/pgtable.h 	pte_val(pte) &= _PAGE_CHG_MASK;
pte               925 arch/s390/include/asm/pgtable.h 	pte_val(pte) |= pgprot_val(newprot);
pte               930 arch/s390/include/asm/pgtable.h 	if ((pte_val(pte) & _PAGE_YOUNG) && (pte_val(pte) & _PAGE_READ))
pte               931 arch/s390/include/asm/pgtable.h 		pte_val(pte) &= ~_PAGE_INVALID;
pte               936 arch/s390/include/asm/pgtable.h 	if ((pte_val(pte) & _PAGE_DIRTY) && (pte_val(pte) & _PAGE_WRITE))
pte               937 arch/s390/include/asm/pgtable.h 		pte_val(pte) &= ~_PAGE_PROTECT;
pte               938 arch/s390/include/asm/pgtable.h 	return pte;
pte               941 arch/s390/include/asm/pgtable.h static inline pte_t pte_wrprotect(pte_t pte)
pte               943 arch/s390/include/asm/pgtable.h 	pte_val(pte) &= ~_PAGE_WRITE;
pte               944 arch/s390/include/asm/pgtable.h 	pte_val(pte) |= _PAGE_PROTECT;
pte               945 arch/s390/include/asm/pgtable.h 	return pte;
pte               948 arch/s390/include/asm/pgtable.h static inline pte_t pte_mkwrite(pte_t pte)
pte               950 arch/s390/include/asm/pgtable.h 	pte_val(pte) |= _PAGE_WRITE;
pte               951 arch/s390/include/asm/pgtable.h 	if (pte_val(pte) & _PAGE_DIRTY)
pte               952 arch/s390/include/asm/pgtable.h 		pte_val(pte) &= ~_PAGE_PROTECT;
pte               953 arch/s390/include/asm/pgtable.h 	return pte;
pte               956 arch/s390/include/asm/pgtable.h static inline pte_t pte_mkclean(pte_t pte)
pte               958 arch/s390/include/asm/pgtable.h 	pte_val(pte) &= ~_PAGE_DIRTY;
pte               959 arch/s390/include/asm/pgtable.h 	pte_val(pte) |= _PAGE_PROTECT;
pte               960 arch/s390/include/asm/pgtable.h 	return pte;
pte               963 arch/s390/include/asm/pgtable.h static inline pte_t pte_mkdirty(pte_t pte)
pte               965 arch/s390/include/asm/pgtable.h 	pte_val(pte) |= _PAGE_DIRTY | _PAGE_SOFT_DIRTY;
pte               966 arch/s390/include/asm/pgtable.h 	if (pte_val(pte) & _PAGE_WRITE)
pte               967 arch/s390/include/asm/pgtable.h 		pte_val(pte) &= ~_PAGE_PROTECT;
pte               968 arch/s390/include/asm/pgtable.h 	return pte;
pte               971 arch/s390/include/asm/pgtable.h static inline pte_t pte_mkold(pte_t pte)
pte               973 arch/s390/include/asm/pgtable.h 	pte_val(pte) &= ~_PAGE_YOUNG;
pte               974 arch/s390/include/asm/pgtable.h 	pte_val(pte) |= _PAGE_INVALID;
pte               975 arch/s390/include/asm/pgtable.h 	return pte;
pte               978 arch/s390/include/asm/pgtable.h static inline pte_t pte_mkyoung(pte_t pte)
pte               980 arch/s390/include/asm/pgtable.h 	pte_val(pte) |= _PAGE_YOUNG;
pte               981 arch/s390/include/asm/pgtable.h 	if (pte_val(pte) & _PAGE_READ)
pte               982 arch/s390/include/asm/pgtable.h 		pte_val(pte) &= ~_PAGE_INVALID;
pte               983 arch/s390/include/asm/pgtable.h 	return pte;
pte               986 arch/s390/include/asm/pgtable.h static inline pte_t pte_mkspecial(pte_t pte)
pte               988 arch/s390/include/asm/pgtable.h 	pte_val(pte) |= _PAGE_SPECIAL;
pte               989 arch/s390/include/asm/pgtable.h 	return pte;
pte               993 arch/s390/include/asm/pgtable.h static inline pte_t pte_mkhuge(pte_t pte)
pte               995 arch/s390/include/asm/pgtable.h 	pte_val(pte) |= _PAGE_LARGE;
pte               996 arch/s390/include/asm/pgtable.h 	return pte;
pte              1063 arch/s390/include/asm/pgtable.h 	pte_t pte = *ptep;
pte              1065 arch/s390/include/asm/pgtable.h 	pte = ptep_xchg_direct(vma->vm_mm, addr, ptep, pte_mkold(pte));
pte              1066 arch/s390/include/asm/pgtable.h 	return pte_young(pte);
pte              1108 arch/s390/include/asm/pgtable.h 		pte_t pte = *ptep;
pte              1110 arch/s390/include/asm/pgtable.h 		return pte;
pte              1119 arch/s390/include/asm/pgtable.h 	pte_t pte = *ptep;
pte              1121 arch/s390/include/asm/pgtable.h 	if (pte_write(pte))
pte              1122 arch/s390/include/asm/pgtable.h 		ptep_xchg_lazy(mm, addr, ptep, pte_wrprotect(pte));
pte              1150 arch/s390/include/asm/pgtable.h 		    pte_t *sptep, pte_t *tptep, pte_t pte);
pte              1279 arch/s390/include/asm/pgtable.h static inline void pte_unmap(pte_t *pte) { }
pte              1654 arch/s390/include/asm/pgtable.h 	pte_t pte;
pte              1656 arch/s390/include/asm/pgtable.h 	pte_val(pte) = _PAGE_INVALID | _PAGE_PROTECT;
pte              1657 arch/s390/include/asm/pgtable.h 	pte_val(pte) |= (offset & __SWP_OFFSET_MASK) << __SWP_OFFSET_SHIFT;
pte              1658 arch/s390/include/asm/pgtable.h 	pte_val(pte) |= (type & __SWP_TYPE_MASK) << __SWP_TYPE_SHIFT;
pte              1659 arch/s390/include/asm/pgtable.h 	return pte;
pte              1677 arch/s390/include/asm/pgtable.h #define __pte_to_swp_entry(pte)	((swp_entry_t) { pte_val(pte) })
pte                64 arch/s390/include/asm/tlb.h static inline void pte_free_tlb(struct mmu_gather *tlb, pgtable_t pte,
pte                76 arch/s390/include/asm/tlb.h 	page_table_free_rcu(tlb, (unsigned long *) pte, address);
pte               620 arch/s390/kvm/gaccess.c 	union page_table_entry pte;
pte               752 arch/s390/kvm/gaccess.c 	if (deref_table(vcpu->kvm, ptr, &pte.val))
pte               754 arch/s390/kvm/gaccess.c 	if (pte.i)
pte               756 arch/s390/kvm/gaccess.c 	if (pte.z)
pte               758 arch/s390/kvm/gaccess.c 	dat_protection |= pte.p;
pte               759 arch/s390/kvm/gaccess.c 	iep_protection = pte.iep;
pte               760 arch/s390/kvm/gaccess.c 	raddr.pfra = pte.pfra;
pte              1168 arch/s390/kvm/gaccess.c 	union page_table_entry pte;
pte              1188 arch/s390/kvm/gaccess.c 		pte.val = pgt + vaddr.px * PAGE_SIZE;
pte              1192 arch/s390/kvm/gaccess.c 		rc = gmap_read_table(sg->parent, pgt + vaddr.px * 8, &pte.val);
pte              1193 arch/s390/kvm/gaccess.c 	if (!rc && pte.i)
pte              1195 arch/s390/kvm/gaccess.c 	if (!rc && pte.z)
pte              1198 arch/s390/kvm/gaccess.c 	pte.p |= dat_protection;
pte              1200 arch/s390/kvm/gaccess.c 		rc = gmap_shadow_page(sg, saddr, __pte(pte.val));
pte               137 arch/s390/mm/dump_pagetables.c 	pte_t *pte;
pte               142 arch/s390/mm/dump_pagetables.c 		pte = pte_offset_kernel(pmd, addr);
pte               143 arch/s390/mm/dump_pagetables.c 		prot = pte_val(*pte) &
pte              1133 arch/s390/mm/gmap.c 	pte_t *ptep, pte;
pte              1143 arch/s390/mm/gmap.c 			pte = *ptep;
pte              1144 arch/s390/mm/gmap.c 			if (pte_present(pte) && (pte_val(pte) & _PAGE_READ)) {
pte              1145 arch/s390/mm/gmap.c 				address = pte_val(pte) & PAGE_MASK;
pte              2105 arch/s390/mm/gmap.c int gmap_shadow_page(struct gmap *sg, unsigned long saddr, pte_t pte)
pte              2117 arch/s390/mm/gmap.c 	prot = (pte_val(pte) & _PAGE_PROTECT) ? PROT_READ : PROT_WRITE;
pte              2125 arch/s390/mm/gmap.c 		paddr = pte_val(pte) & PAGE_MASK;
pte              2146 arch/s390/mm/gmap.c 			rc = ptep_shadow_pte(sg->mm, saddr, sptep, tptep, pte);
pte              2236 arch/s390/mm/gmap.c 		 pte_t *pte, unsigned long bits)
pte              2242 arch/s390/mm/gmap.c 	offset = ((unsigned long) pte) & (255 * sizeof(pte_t));
pte              2560 arch/s390/mm/gmap.c static int __s390_enable_skey_pte(pte_t *pte, unsigned long addr,
pte              2564 arch/s390/mm/gmap.c 	ptep_zap_key(walk->mm, addr, pte);
pte              2568 arch/s390/mm/gmap.c static int __s390_enable_skey_hugetlb(pte_t *pte, unsigned long addr,
pte              2572 arch/s390/mm/gmap.c 	pmd_t *pmd = (pmd_t *)pte;
pte              2630 arch/s390/mm/gmap.c static int __s390_reset_cmma(pte_t *pte, unsigned long addr,
pte              2633 arch/s390/mm/gmap.c 	ptep_zap_unused(walk->mm, addr, pte, 1);
pte                24 arch/s390/mm/hugetlbpage.c static inline unsigned long __pte_to_rste(pte_t pte)
pte                48 arch/s390/mm/hugetlbpage.c 	if (pte_present(pte)) {
pte                49 arch/s390/mm/hugetlbpage.c 		rste = pte_val(pte) & PAGE_MASK;
pte                50 arch/s390/mm/hugetlbpage.c 		rste |= move_set_bit(pte_val(pte), _PAGE_READ,
pte                52 arch/s390/mm/hugetlbpage.c 		rste |= move_set_bit(pte_val(pte), _PAGE_WRITE,
pte                54 arch/s390/mm/hugetlbpage.c 		rste |= move_set_bit(pte_val(pte), _PAGE_INVALID,
pte                56 arch/s390/mm/hugetlbpage.c 		rste |= move_set_bit(pte_val(pte), _PAGE_PROTECT,
pte                58 arch/s390/mm/hugetlbpage.c 		rste |= move_set_bit(pte_val(pte), _PAGE_DIRTY,
pte                60 arch/s390/mm/hugetlbpage.c 		rste |= move_set_bit(pte_val(pte), _PAGE_YOUNG,
pte                63 arch/s390/mm/hugetlbpage.c 		rste |= move_set_bit(pte_val(pte), _PAGE_SOFT_DIRTY,
pte                66 arch/s390/mm/hugetlbpage.c 		rste |= move_set_bit(pte_val(pte), _PAGE_NOEXEC,
pte                76 arch/s390/mm/hugetlbpage.c 	pte_t pte;
pte               104 arch/s390/mm/hugetlbpage.c 		pte_val(pte) = rste & _SEGMENT_ENTRY_ORIGIN_LARGE;
pte               105 arch/s390/mm/hugetlbpage.c 		pte_val(pte) |= _PAGE_LARGE | _PAGE_PRESENT;
pte               106 arch/s390/mm/hugetlbpage.c 		pte_val(pte) |= move_set_bit(rste, _SEGMENT_ENTRY_READ,
pte               108 arch/s390/mm/hugetlbpage.c 		pte_val(pte) |= move_set_bit(rste, _SEGMENT_ENTRY_WRITE,
pte               110 arch/s390/mm/hugetlbpage.c 		pte_val(pte) |= move_set_bit(rste, _SEGMENT_ENTRY_INVALID,
pte               112 arch/s390/mm/hugetlbpage.c 		pte_val(pte) |= move_set_bit(rste, _SEGMENT_ENTRY_PROTECT,
pte               114 arch/s390/mm/hugetlbpage.c 		pte_val(pte) |= move_set_bit(rste, _SEGMENT_ENTRY_DIRTY,
pte               116 arch/s390/mm/hugetlbpage.c 		pte_val(pte) |= move_set_bit(rste, _SEGMENT_ENTRY_YOUNG,
pte               119 arch/s390/mm/hugetlbpage.c 		pte_val(pte) |= move_set_bit(rste, _SEGMENT_ENTRY_SOFT_DIRTY,
pte               122 arch/s390/mm/hugetlbpage.c 		pte_val(pte) |= move_set_bit(rste, _SEGMENT_ENTRY_NOEXEC,
pte               125 arch/s390/mm/hugetlbpage.c 		pte_val(pte) = _PAGE_INVALID;
pte               126 arch/s390/mm/hugetlbpage.c 	return pte;
pte               153 arch/s390/mm/hugetlbpage.c 		     pte_t *ptep, pte_t pte)
pte               157 arch/s390/mm/hugetlbpage.c 	rste = __pte_to_rste(pte);
pte               163 arch/s390/mm/hugetlbpage.c 		if (likely(pte_present(pte)))
pte               166 arch/s390/mm/hugetlbpage.c 	} else if (likely(pte_present(pte)))
pte               181 arch/s390/mm/hugetlbpage.c 	pte_t pte = huge_ptep_get(ptep);
pte               189 arch/s390/mm/hugetlbpage.c 	return pte;
pte                67 arch/s390/mm/kasan_init.c 	pte_t *pte;
pte                73 arch/s390/mm/kasan_init.c 		pte = pte_leftover + _PAGE_TABLE_SIZE;
pte                75 arch/s390/mm/kasan_init.c 		pte = pte_leftover;
pte                78 arch/s390/mm/kasan_init.c 	memset64((u64 *)pte, _PAGE_INVALID, PTRS_PER_PTE);
pte                79 arch/s390/mm/kasan_init.c 	return pte;
pte               322 arch/s390/mm/pageattr.c static void ipte_range(pte_t *pte, unsigned long address, int nr)
pte               327 arch/s390/mm/pageattr.c 		__ptep_ipte_range(address, nr - 1, pte, IPTE_GLOBAL);
pte               331 arch/s390/mm/pageattr.c 		__ptep_ipte(address, pte, 0, 0, IPTE_GLOBAL);
pte               333 arch/s390/mm/pageattr.c 		pte++;
pte               345 arch/s390/mm/pageattr.c 	pte_t *pte;
pte               353 arch/s390/mm/pageattr.c 		pte = pte_offset_kernel(pmd, address);
pte               354 arch/s390/mm/pageattr.c 		nr = (unsigned long)pte >> ilog2(sizeof(long));
pte               359 arch/s390/mm/pageattr.c 				pte_val(*pte) &= ~_PAGE_INVALID;
pte               361 arch/s390/mm/pageattr.c 				pte++;
pte               364 arch/s390/mm/pageattr.c 			ipte_range(pte, address, nr);
pte               394 arch/s390/mm/pgalloc.c 	unsigned long *pte, next;
pte               398 arch/s390/mm/pgalloc.c 	pte = (unsigned long *) origin;
pte               399 arch/s390/mm/pgalloc.c 	pte += (addr & _PAGE_INDEX) >> _PAGE_SHIFT;
pte               402 arch/s390/mm/pgalloc.c 		*pte = base_lra(addr);
pte               403 arch/s390/mm/pgalloc.c 	} while (pte++, addr = next, addr < end);
pte               155 arch/s390/mm/pgtable.c static inline pgste_t pgste_update_all(pte_t pte, pgste_t pgste,
pte               161 arch/s390/mm/pgtable.c 	if (!mm_uses_skeys(mm) || pte_val(pte) & _PAGE_INVALID)
pte               163 arch/s390/mm/pgtable.c 	address = pte_val(pte) & PAGE_MASK;
pte               324 arch/s390/mm/pgtable.c 			     pte_t *ptep, pte_t old_pte, pte_t pte)
pte               330 arch/s390/mm/pgtable.c 		pte_val(pte) &= ~_PAGE_NOEXEC;
pte               333 arch/s390/mm/pgtable.c 		pgste_set_key(ptep, pgste, pte, mm);
pte               334 arch/s390/mm/pgtable.c 		pgste = pgste_set_pte(ptep, pgste, pte);
pte               337 arch/s390/mm/pgtable.c 		*ptep = pte;
pte               631 arch/s390/mm/pgtable.c 		    pte_t *sptep, pte_t *tptep, pte_t pte)
pte               643 arch/s390/mm/pgtable.c 	      !(pte_val(pte) & _PAGE_PROTECT))) {
pte               647 arch/s390/mm/pgtable.c 				(pte_val(pte) & _PAGE_PROTECT);
pte               688 arch/s390/mm/pgtable.c 	pte_t pte;
pte               694 arch/s390/mm/pgtable.c 	pte = *ptep;
pte               695 arch/s390/mm/pgtable.c 	if (!reset && pte_swap(pte) &&
pte               698 arch/s390/mm/pgtable.c 		ptep_zap_swap_entry(mm, pte_to_swp_entry(pte));
pte               731 arch/s390/mm/pgtable.c 	pte_t pte;
pte               738 arch/s390/mm/pgtable.c 	pte = *ptep;
pte               739 arch/s390/mm/pgtable.c 	if (dirty && (pte_val(pte) & _PAGE_PRESENT)) {
pte               743 arch/s390/mm/pgtable.c 		if (MACHINE_HAS_ESOP || !(pte_val(pte) & _PAGE_WRITE))
pte               744 arch/s390/mm/pgtable.c 			pte_val(pte) |= _PAGE_PROTECT;
pte               746 arch/s390/mm/pgtable.c 			pte_val(pte) |= _PAGE_INVALID;
pte               747 arch/s390/mm/pgtable.c 		*ptep = pte;
pte                54 arch/s390/mm/vmem.c 	pte_t *pte;
pte                57 arch/s390/mm/vmem.c 		pte = (pte_t *) page_table_alloc(&init_mm);
pte                59 arch/s390/mm/vmem.c 		pte = (pte_t *) memblock_phys_alloc(size, size);
pte                60 arch/s390/mm/vmem.c 	if (!pte)
pte                62 arch/s390/mm/vmem.c 	memset64((u64 *)pte, _PAGE_INVALID, PTRS_PER_PTE);
pte                63 arch/s390/mm/vmem.c 	return pte;
pte                18 arch/sh/include/asm/pgalloc.h 				       pte_t *pte)
pte                20 arch/sh/include/asm/pgalloc.h 	set_pmd(pmd, __pmd((unsigned long)pte));
pte                24 arch/sh/include/asm/pgalloc.h 				pgtable_t pte)
pte                26 arch/sh/include/asm/pgalloc.h 	set_pmd(pmd, __pmd((unsigned long)page_address(pte)));
pte                30 arch/sh/include/asm/pgalloc.h #define __pte_free_tlb(tlb,pte,addr)			\
pte                32 arch/sh/include/asm/pgalloc.h 	pgtable_pte_page_dtor(pte);			\
pte                33 arch/sh/include/asm/pgalloc.h 	tlb_remove_page((tlb), (pte));			\
pte               130 arch/sh/include/asm/pgtable.h 			   unsigned long address, pte_t pte);
pte               132 arch/sh/include/asm/pgtable.h 			 unsigned long address, pte_t pte);
pte               137 arch/sh/include/asm/pgtable.h 	pte_t pte = *ptep;
pte               138 arch/sh/include/asm/pgtable.h 	__update_cache(vma, address, pte);
pte               139 arch/sh/include/asm/pgtable.h 	__update_tlb(vma, address, pte);
pte               147 arch/sh/include/asm/pgtable.h static inline bool __pte_access_permitted(pte_t pte, u64 prot)
pte               149 arch/sh/include/asm/pgtable.h 	return (pte_val(pte) & (prot | _PAGE_SPECIAL)) == prot;
pte               153 arch/sh/include/asm/pgtable.h static inline bool pte_access_permitted(pte_t pte, bool write)
pte               160 arch/sh/include/asm/pgtable.h 	return __pte_access_permitted(pte, prot);
pte               163 arch/sh/include/asm/pgtable.h static inline bool pte_access_permitted(pte_t pte, bool write)
pte               169 arch/sh/include/asm/pgtable.h 	return __pte_access_permitted(pte, prot);
pte               172 arch/sh/include/asm/pgtable.h static inline bool pte_access_permitted(pte_t pte, bool write)
pte               178 arch/sh/include/asm/pgtable.h 	return __pte_access_permitted(pte, prot);
pte               300 arch/sh/include/asm/pgtable_32.h static inline void set_pte(pte_t *ptep, pte_t pte)
pte               302 arch/sh/include/asm/pgtable_32.h 	ptep->pte_high = pte.pte_high;
pte               304 arch/sh/include/asm/pgtable_32.h 	ptep->pte_low = pte.pte_low;
pte               340 arch/sh/include/asm/pgtable_32.h #define pte_not_present(pte)	(!((pte).pte_low & _PAGE_PRESENT))
pte               341 arch/sh/include/asm/pgtable_32.h #define pte_dirty(pte)		((pte).pte_low & _PAGE_DIRTY)
pte               342 arch/sh/include/asm/pgtable_32.h #define pte_young(pte)		((pte).pte_low & _PAGE_ACCESSED)
pte               343 arch/sh/include/asm/pgtable_32.h #define pte_special(pte)	((pte).pte_low & _PAGE_SPECIAL)
pte               346 arch/sh/include/asm/pgtable_32.h #define pte_write(pte) \
pte               347 arch/sh/include/asm/pgtable_32.h 	((pte).pte_high & (_PAGE_EXT_USER_WRITE | _PAGE_EXT_KERN_WRITE))
pte               349 arch/sh/include/asm/pgtable_32.h #define pte_write(pte)		((pte).pte_low & _PAGE_RW)
pte               353 arch/sh/include/asm/pgtable_32.h static inline pte_t pte_##fn(pte_t pte) { pte.pte_##h op; return pte; }
pte               392 arch/sh/include/asm/pgtable_32.h static inline pte_t pte_modify(pte_t pte, pgprot_t newprot)
pte               394 arch/sh/include/asm/pgtable_32.h 	pte.pte_low &= _PAGE_CHG_MASK;
pte               395 arch/sh/include/asm/pgtable_32.h 	pte.pte_low |= pgprot_val(newprot);
pte               398 arch/sh/include/asm/pgtable_32.h 	pte.pte_high |= pgprot_val(newprot) >> 32;
pte               401 arch/sh/include/asm/pgtable_32.h 	return pte;
pte               425 arch/sh/include/asm/pgtable_32.h #define pte_unmap(pte)		do { } while (0)
pte               464 arch/sh/include/asm/pgtable_32.h #define __pte_to_swp_entry(pte)		((swp_entry_t){ (pte).pte_high })
pte               472 arch/sh/include/asm/pgtable_32.h #define __pte_to_swp_entry(pte)		((swp_entry_t) { pte_val(pte) >> 1 })
pte                84 arch/sh/include/asm/pgtable_64.h #define pte_unmap(pte)		do { } while (0)
pte               257 arch/sh/include/asm/pgtable_64.h static inline int pte_dirty(pte_t pte)  { return pte_val(pte) & _PAGE_DIRTY; }
pte               258 arch/sh/include/asm/pgtable_64.h static inline int pte_young(pte_t pte)  { return pte_val(pte) & _PAGE_ACCESSED; }
pte               259 arch/sh/include/asm/pgtable_64.h static inline int pte_write(pte_t pte)  { return pte_val(pte) & _PAGE_WRITE; }
pte               260 arch/sh/include/asm/pgtable_64.h static inline int pte_special(pte_t pte){ return pte_val(pte) & _PAGE_SPECIAL; }
pte               262 arch/sh/include/asm/pgtable_64.h static inline pte_t pte_wrprotect(pte_t pte)	{ set_pte(&pte, __pte(pte_val(pte) & ~_PAGE_WRITE)); return pte; }
pte               263 arch/sh/include/asm/pgtable_64.h static inline pte_t pte_mkclean(pte_t pte)	{ set_pte(&pte, __pte(pte_val(pte) & ~_PAGE_DIRTY)); return pte; }
pte               264 arch/sh/include/asm/pgtable_64.h static inline pte_t pte_mkold(pte_t pte)	{ set_pte(&pte, __pte(pte_val(pte) & ~_PAGE_ACCESSED)); return pte; }
pte               265 arch/sh/include/asm/pgtable_64.h static inline pte_t pte_mkwrite(pte_t pte)	{ set_pte(&pte, __pte(pte_val(pte) | _PAGE_WRITE)); return pte; }
pte               266 arch/sh/include/asm/pgtable_64.h static inline pte_t pte_mkdirty(pte_t pte)	{ set_pte(&pte, __pte(pte_val(pte) | _PAGE_DIRTY)); return pte; }
pte               267 arch/sh/include/asm/pgtable_64.h static inline pte_t pte_mkyoung(pte_t pte)	{ set_pte(&pte, __pte(pte_val(pte) | _PAGE_ACCESSED)); return pte; }
pte               268 arch/sh/include/asm/pgtable_64.h static inline pte_t pte_mkhuge(pte_t pte)	{ set_pte(&pte, __pte(pte_val(pte) | _PAGE_SZHUGE)); return pte; }
pte               269 arch/sh/include/asm/pgtable_64.h static inline pte_t pte_mkspecial(pte_t pte)	{ set_pte(&pte, __pte(pte_val(pte) | _PAGE_SPECIAL)); return pte; }
pte               292 arch/sh/include/asm/pgtable_64.h static inline pte_t pte_modify(pte_t pte, pgprot_t newprot)
pte               293 arch/sh/include/asm/pgtable_64.h { set_pte(&pte, __pte((pte_val(pte) & _PAGE_CHG_MASK) | pgprot_val(newprot))); return pte; }
pte               299 arch/sh/include/asm/pgtable_64.h #define __pte_to_swp_entry(pte)		((swp_entry_t) { pte_val(pte) })
pte                22 arch/sh/include/asm/tlb.h 				  unsigned long addr, pte_t pte)
pte                19 arch/sh/include/cpu-sh5/cpu/mmu_context.h extern void __do_tlb_refill(unsigned long address, unsigned long long is_text_not_data, pte_t *pte);
pte               214 arch/sh/mm/cache-sh4.c 	pte_t *pte;
pte               229 arch/sh/mm/cache-sh4.c 	pte = pte_offset_kernel(pmd, address);
pte               232 arch/sh/mm/cache-sh4.c 	if (!(pte_val(*pte) & _PAGE_PRESENT))
pte               388 arch/sh/mm/cache-sh5.c 	pte_t *pte;
pte               408 arch/sh/mm/cache-sh5.c 	pte = pte_offset_map_lock(mm, pmd, addr, &ptl);
pte               410 arch/sh/mm/cache-sh5.c 		entry = *pte;
pte               415 arch/sh/mm/cache-sh5.c 	} while (pte++, addr += PAGE_SIZE, addr != end);
pte               416 arch/sh/mm/cache-sh5.c 	pte_unmap_unlock(pte - 1, ptl);
pte               137 arch/sh/mm/cache.c 		    unsigned long address, pte_t pte)
pte               140 arch/sh/mm/cache.c 	unsigned long pfn = pte_pfn(pte);
pte                58 arch/sh/mm/fault.c 		pte_t *pte;
pte                98 arch/sh/mm/fault.c 		pte = pte_offset_kernel(pmd, addr);
pte                99 arch/sh/mm/fault.c 		printk(", *pte=%0*Lx", (u32)(sizeof(*pte) * 2),
pte               100 arch/sh/mm/fault.c 		       (u64)pte_val(*pte));
pte                31 arch/sh/mm/hugetlbpage.c 	pte_t *pte = NULL;
pte                39 arch/sh/mm/hugetlbpage.c 				pte = pte_alloc_map(mm, pmd, addr);
pte                43 arch/sh/mm/hugetlbpage.c 	return pte;
pte                52 arch/sh/mm/hugetlbpage.c 	pte_t *pte = NULL;
pte                60 arch/sh/mm/hugetlbpage.c 				pte = pte_offset_map(pmd, addr);
pte                64 arch/sh/mm/hugetlbpage.c 	return pte;
pte                74 arch/sh/mm/init.c 	pte_t *pte;
pte                76 arch/sh/mm/init.c 	pte = __get_pte_phys(addr);
pte                77 arch/sh/mm/init.c 	if (!pte_none(*pte)) {
pte                78 arch/sh/mm/init.c 		pte_ERROR(*pte);
pte                82 arch/sh/mm/init.c 	set_pte(pte, pfn_pte(phys >> PAGE_SHIFT, prot));
pte                86 arch/sh/mm/init.c 		tlb_wire_entry(NULL, addr, *pte);
pte                91 arch/sh/mm/init.c 	pte_t *pte;
pte                93 arch/sh/mm/init.c 	pte = __get_pte_phys(addr);
pte                98 arch/sh/mm/init.c 	set_pte(pte, pfn_pte(0, __pgprot(0)));
pte               145 arch/sh/mm/init.c 		pte_t *pte;
pte               147 arch/sh/mm/init.c 		pte = memblock_alloc(PAGE_SIZE, PAGE_SIZE);
pte               148 arch/sh/mm/init.c 		if (!pte)
pte               151 arch/sh/mm/init.c 		pmd_populate_kernel(&init_mm, pmd, pte);
pte               152 arch/sh/mm/init.c 		BUG_ON(pte != pte_offset_kernel(pmd, 0));
pte               158 arch/sh/mm/init.c static pte_t * __init page_table_kmap_check(pte_t *pte, pmd_t *pmd,
pte               161 arch/sh/mm/init.c 	return pte;
pte               170 arch/sh/mm/init.c 	pte_t *pte = NULL;
pte               188 arch/sh/mm/init.c 				pte = page_table_kmap_check(one_page_table_init(pmd),
pte               189 arch/sh/mm/init.c 							    pmd, vaddr, pte);
pte                73 arch/sh/mm/nommu.c void __update_tlb(struct vm_area_struct *vma, unsigned long address, pte_t pte)
pte                18 arch/sh/mm/tlb-pteaex.c void __update_tlb(struct vm_area_struct *vma, unsigned long address, pte_t pte)
pte                37 arch/sh/mm/tlb-pteaex.c 	pteval = pte.pte_low;
pte                47 arch/sh/mm/tlb-pteaex.c 	__raw_writel(pte.pte_high, MMU_PTEA);
pte                28 arch/sh/mm/tlb-sh3.c void __update_tlb(struct vm_area_struct *vma, unsigned long address, pte_t pte)
pte                44 arch/sh/mm/tlb-sh3.c 	pteval = pte_val(pte);
pte                16 arch/sh/mm/tlb-sh4.c void __update_tlb(struct vm_area_struct *vma, unsigned long address, pte_t pte)
pte                32 arch/sh/mm/tlb-sh4.c 	pteval = pte.pte_low;
pte                42 arch/sh/mm/tlb-sh4.c 	__raw_writel(pte.pte_high, MMU_PTEA);
pte               150 arch/sh/mm/tlb-sh5.c void tlb_wire_entry(struct vm_area_struct *vma, unsigned long addr, pte_t pte)
pte               162 arch/sh/mm/tlb-sh5.c 	paddr = pte_val(pte) & _PAGE_FLAGS_HARDWARE_MASK;
pte               186 arch/sh/mm/tlb-sh5.c void __update_tlb(struct vm_area_struct *vma, unsigned long address, pte_t pte)
pte               195 arch/sh/mm/tlb-sh5.c 	ptel = pte.pte_low;
pte                20 arch/sh/mm/tlb-urb.c void tlb_wire_entry(struct vm_area_struct *vma, unsigned long addr, pte_t pte)
pte                47 arch/sh/mm/tlb-urb.c 	__update_tlb(vma, addr, pte);
pte                28 arch/sh/mm/tlbex_32.c 	pte_t *pte;
pte                51 arch/sh/mm/tlbex_32.c 	pte = pte_offset_kernel(pmd, address);
pte                52 arch/sh/mm/tlbex_32.c 	entry = *pte;
pte                62 arch/sh/mm/tlbex_32.c 	set_pte(pte, entry);
pte                75 arch/sh/mm/tlbex_32.c 	update_mmu_cache(NULL, address, pte);
pte                49 arch/sh/mm/tlbex_64.c 	pte_t *pte;
pte                69 arch/sh/mm/tlbex_64.c 	pte = pte_offset_kernel(pmd, address);
pte                70 arch/sh/mm/tlbex_64.c 	entry = *pte;
pte                83 arch/sh/mm/tlbex_64.c 	update_mmu_cache(NULL, address, pte);
pte                17 arch/sparc/include/asm/hugetlb.h 		     pte_t *ptep, pte_t pte);
pte                46 arch/sparc/include/asm/hugetlb.h 					     pte_t pte, int dirty)
pte                48 arch/sparc/include/asm/hugetlb.h 	int changed = !pte_same(*ptep, pte);
pte                50 arch/sparc/include/asm/hugetlb.h 		set_huge_pte_at(vma->vm_mm, addr, ptep, pte);
pte                68 arch/sparc/include/asm/mmu_64.h 	unsigned long pte;
pte                71 arch/sparc/include/asm/mmu_64.h void __tsb_insert(unsigned long ent, unsigned long tag, unsigned long pte);
pte                55 arch/sparc/include/asm/page_32.h typedef struct { unsigned long pte; } pte_t;
pte                63 arch/sparc/include/asm/page_32.h #define pte_val(x)	((x).pte)
pte                69 arch/sparc/include/asm/page_64.h typedef struct { unsigned long pte; } pte_t;
pte                76 arch/sparc/include/asm/page_64.h #define pte_val(x)	((x).pte)
pte                67 arch/sparc/include/asm/pgalloc_32.h static inline void free_pte_fast(pte_t *pte)
pte                69 arch/sparc/include/asm/pgalloc_32.h 	srmmu_free_nocache(pte, PTE_SIZE);
pte                72 arch/sparc/include/asm/pgalloc_32.h #define pte_free_kernel(mm, pte)	free_pte_fast(pte)
pte                74 arch/sparc/include/asm/pgalloc_32.h void pte_free(struct mm_struct * mm, pgtable_t pte);
pte                75 arch/sparc/include/asm/pgalloc_32.h #define __pte_free_tlb(tlb, pte, addr)	pte_free((tlb)->mm, pte)
pte                65 arch/sparc/include/asm/pgalloc_64.h void pte_free_kernel(struct mm_struct *mm, pte_t *pte);
pte               103 arch/sparc/include/asm/pgalloc_64.h static inline void __pte_free_tlb(struct mmu_gather *tlb, pte_t *pte,
pte               106 arch/sparc/include/asm/pgalloc_64.h 	pgtable_free_tlb(tlb, pte, true);
pte               145 arch/sparc/include/asm/pgtable_32.h static inline int pte_present(pte_t pte)
pte               147 arch/sparc/include/asm/pgtable_32.h 	return ((pte_val(pte) & SRMMU_ET_MASK) == SRMMU_ET_PTE);
pte               150 arch/sparc/include/asm/pgtable_32.h static inline int pte_none(pte_t pte)
pte               152 arch/sparc/include/asm/pgtable_32.h 	return !pte_val(pte);
pte               211 arch/sparc/include/asm/pgtable_32.h static inline int pte_write(pte_t pte)
pte               213 arch/sparc/include/asm/pgtable_32.h 	return pte_val(pte) & SRMMU_WRITE;
pte               216 arch/sparc/include/asm/pgtable_32.h static inline int pte_dirty(pte_t pte)
pte               218 arch/sparc/include/asm/pgtable_32.h 	return pte_val(pte) & SRMMU_DIRTY;
pte               221 arch/sparc/include/asm/pgtable_32.h static inline int pte_young(pte_t pte)
pte               223 arch/sparc/include/asm/pgtable_32.h 	return pte_val(pte) & SRMMU_REF;
pte               226 arch/sparc/include/asm/pgtable_32.h static inline int pte_special(pte_t pte)
pte               231 arch/sparc/include/asm/pgtable_32.h static inline pte_t pte_wrprotect(pte_t pte)
pte               233 arch/sparc/include/asm/pgtable_32.h 	return __pte(pte_val(pte) & ~SRMMU_WRITE);
pte               236 arch/sparc/include/asm/pgtable_32.h static inline pte_t pte_mkclean(pte_t pte)
pte               238 arch/sparc/include/asm/pgtable_32.h 	return __pte(pte_val(pte) & ~SRMMU_DIRTY);
pte               241 arch/sparc/include/asm/pgtable_32.h static inline pte_t pte_mkold(pte_t pte)
pte               243 arch/sparc/include/asm/pgtable_32.h 	return __pte(pte_val(pte) & ~SRMMU_REF);
pte               246 arch/sparc/include/asm/pgtable_32.h static inline pte_t pte_mkwrite(pte_t pte)
pte               248 arch/sparc/include/asm/pgtable_32.h 	return __pte(pte_val(pte) | SRMMU_WRITE);
pte               251 arch/sparc/include/asm/pgtable_32.h static inline pte_t pte_mkdirty(pte_t pte)
pte               253 arch/sparc/include/asm/pgtable_32.h 	return __pte(pte_val(pte) | SRMMU_DIRTY);
pte               256 arch/sparc/include/asm/pgtable_32.h static inline pte_t pte_mkyoung(pte_t pte)
pte               258 arch/sparc/include/asm/pgtable_32.h 	return __pte(pte_val(pte) | SRMMU_REF);
pte               261 arch/sparc/include/asm/pgtable_32.h #define pte_mkspecial(pte)    (pte)
pte               265 arch/sparc/include/asm/pgtable_32.h static inline unsigned long pte_pfn(pte_t pte)
pte               267 arch/sparc/include/asm/pgtable_32.h 	if (srmmu_device_memory(pte_val(pte))) {
pte               275 arch/sparc/include/asm/pgtable_32.h 	return (pte_val(pte) & SRMMU_PTE_PMASK) >> (PAGE_SHIFT-4);
pte               278 arch/sparc/include/asm/pgtable_32.h #define pte_page(pte)	pfn_to_page(pte_pfn(pte))
pte               306 arch/sparc/include/asm/pgtable_32.h static pte_t pte_modify(pte_t pte, pgprot_t newprot) __attribute_const__;
pte               307 arch/sparc/include/asm/pgtable_32.h static inline pte_t pte_modify(pte_t pte, pgprot_t newprot)
pte               309 arch/sparc/include/asm/pgtable_32.h 	return __pte((pte_val(pte) & SRMMU_CHG_MASK) |
pte               335 arch/sparc/include/asm/pgtable_32.h #define pte_unmap(pte)		do{}while(0)
pte               369 arch/sparc/include/asm/pgtable_32.h #define __pte_to_swp_entry(pte)		((swp_entry_t) { pte_val(pte) })
pte               251 arch/sparc/include/asm/pgtable_64.h 	pte_t pte = pfn_pte(page_nr, pgprot);
pte               253 arch/sparc/include/asm/pgtable_64.h 	return __pmd(pte_val(pte));
pte               259 arch/sparc/include/asm/pgtable_64.h static inline unsigned long pte_pfn(pte_t pte)
pte               272 arch/sparc/include/asm/pgtable_64.h 	: "r" (pte_val(pte)),
pte               280 arch/sparc/include/asm/pgtable_64.h static inline pte_t pte_modify(pte_t pte, pgprot_t prot)
pte               335 arch/sparc/include/asm/pgtable_64.h 	return __pte((pte_val(pte) & mask) | (pgprot_val(prot) & ~mask));
pte               341 arch/sparc/include/asm/pgtable_64.h 	pte_t pte = __pte(pmd_val(pmd));
pte               343 arch/sparc/include/asm/pgtable_64.h 	pte = pte_modify(pte, newprot);
pte               345 arch/sparc/include/asm/pgtable_64.h 	return __pmd(pte_val(pte));
pte               401 arch/sparc/include/asm/pgtable_64.h static inline pte_t pte_mkhuge(pte_t pte)
pte               403 arch/sparc/include/asm/pgtable_64.h 	return __pte(pte_val(pte) | __pte_default_huge_mask());
pte               406 arch/sparc/include/asm/pgtable_64.h static inline bool is_default_hugetlb_pte(pte_t pte)
pte               410 arch/sparc/include/asm/pgtable_64.h 	return (pte_val(pte) & mask) == mask;
pte               426 arch/sparc/include/asm/pgtable_64.h 	pte_t pte = __pte(pmd_val(pmd));
pte               428 arch/sparc/include/asm/pgtable_64.h 	pte = pte_mkhuge(pte);
pte               429 arch/sparc/include/asm/pgtable_64.h 	pte_val(pte) |= _PAGE_PMD_HUGE;
pte               431 arch/sparc/include/asm/pgtable_64.h 	return __pmd(pte_val(pte));
pte               435 arch/sparc/include/asm/pgtable_64.h static inline bool is_hugetlb_pte(pte_t pte)
pte               441 arch/sparc/include/asm/pgtable_64.h static inline pte_t pte_mkdirty(pte_t pte)
pte               443 arch/sparc/include/asm/pgtable_64.h 	unsigned long val = pte_val(pte), tmp;
pte               465 arch/sparc/include/asm/pgtable_64.h static inline pte_t pte_mkclean(pte_t pte)
pte               467 arch/sparc/include/asm/pgtable_64.h 	unsigned long val = pte_val(pte), tmp;
pte               489 arch/sparc/include/asm/pgtable_64.h static inline pte_t pte_mkwrite(pte_t pte)
pte               491 arch/sparc/include/asm/pgtable_64.h 	unsigned long val = pte_val(pte), mask;
pte               507 arch/sparc/include/asm/pgtable_64.h static inline pte_t pte_wrprotect(pte_t pte)
pte               509 arch/sparc/include/asm/pgtable_64.h 	unsigned long val = pte_val(pte), tmp;
pte               531 arch/sparc/include/asm/pgtable_64.h static inline pte_t pte_mkold(pte_t pte)
pte               548 arch/sparc/include/asm/pgtable_64.h 	return __pte(pte_val(pte) & ~mask);
pte               551 arch/sparc/include/asm/pgtable_64.h static inline pte_t pte_mkyoung(pte_t pte)
pte               568 arch/sparc/include/asm/pgtable_64.h 	return __pte(pte_val(pte) | mask);
pte               571 arch/sparc/include/asm/pgtable_64.h static inline pte_t pte_mkspecial(pte_t pte)
pte               573 arch/sparc/include/asm/pgtable_64.h 	pte_val(pte) |= _PAGE_SPECIAL;
pte               574 arch/sparc/include/asm/pgtable_64.h 	return pte;
pte               577 arch/sparc/include/asm/pgtable_64.h static inline pte_t pte_mkmcd(pte_t pte)
pte               579 arch/sparc/include/asm/pgtable_64.h 	pte_val(pte) |= _PAGE_MCD_4V;
pte               580 arch/sparc/include/asm/pgtable_64.h 	return pte;
pte               583 arch/sparc/include/asm/pgtable_64.h static inline pte_t pte_mknotmcd(pte_t pte)
pte               585 arch/sparc/include/asm/pgtable_64.h 	pte_val(pte) &= ~_PAGE_MCD_4V;
pte               586 arch/sparc/include/asm/pgtable_64.h 	return pte;
pte               589 arch/sparc/include/asm/pgtable_64.h static inline unsigned long pte_young(pte_t pte)
pte               604 arch/sparc/include/asm/pgtable_64.h 	return (pte_val(pte) & mask);
pte               607 arch/sparc/include/asm/pgtable_64.h static inline unsigned long pte_dirty(pte_t pte)
pte               622 arch/sparc/include/asm/pgtable_64.h 	return (pte_val(pte) & mask);
pte               625 arch/sparc/include/asm/pgtable_64.h static inline unsigned long pte_write(pte_t pte)
pte               640 arch/sparc/include/asm/pgtable_64.h 	return (pte_val(pte) & mask);
pte               643 arch/sparc/include/asm/pgtable_64.h static inline unsigned long pte_exec(pte_t pte)
pte               656 arch/sparc/include/asm/pgtable_64.h 	return (pte_val(pte) & mask);
pte               659 arch/sparc/include/asm/pgtable_64.h static inline unsigned long pte_present(pte_t pte)
pte               661 arch/sparc/include/asm/pgtable_64.h 	unsigned long val = pte_val(pte);
pte               681 arch/sparc/include/asm/pgtable_64.h static inline unsigned long pte_special(pte_t pte)
pte               683 arch/sparc/include/asm/pgtable_64.h 	return pte_val(pte) & _PAGE_SPECIAL;
pte               688 arch/sparc/include/asm/pgtable_64.h 	pte_t pte = __pte(pmd_val(pmd));
pte               690 arch/sparc/include/asm/pgtable_64.h 	return pte_val(pte) & _PAGE_PMD_HUGE;
pte               695 arch/sparc/include/asm/pgtable_64.h 	pte_t pte = __pte(pmd_val(pmd));
pte               697 arch/sparc/include/asm/pgtable_64.h 	return pte_pfn(pte);
pte               703 arch/sparc/include/asm/pgtable_64.h 	pte_t pte = __pte(pmd_val(pmd));
pte               705 arch/sparc/include/asm/pgtable_64.h 	return pte_write(pte);
pte               713 arch/sparc/include/asm/pgtable_64.h 	pte_t pte = __pte(pmd_val(pmd));
pte               715 arch/sparc/include/asm/pgtable_64.h 	return pte_dirty(pte);
pte               720 arch/sparc/include/asm/pgtable_64.h 	pte_t pte = __pte(pmd_val(pmd));
pte               722 arch/sparc/include/asm/pgtable_64.h 	return pte_young(pte);
pte               727 arch/sparc/include/asm/pgtable_64.h 	pte_t pte = __pte(pmd_val(pmd));
pte               729 arch/sparc/include/asm/pgtable_64.h 	return pte_val(pte) & _PAGE_PMD_HUGE;
pte               734 arch/sparc/include/asm/pgtable_64.h 	pte_t pte = __pte(pmd_val(pmd));
pte               736 arch/sparc/include/asm/pgtable_64.h 	pte = pte_mkold(pte);
pte               738 arch/sparc/include/asm/pgtable_64.h 	return __pmd(pte_val(pte));
pte               743 arch/sparc/include/asm/pgtable_64.h 	pte_t pte = __pte(pmd_val(pmd));
pte               745 arch/sparc/include/asm/pgtable_64.h 	pte = pte_wrprotect(pte);
pte               747 arch/sparc/include/asm/pgtable_64.h 	return __pmd(pte_val(pte));
pte               752 arch/sparc/include/asm/pgtable_64.h 	pte_t pte = __pte(pmd_val(pmd));
pte               754 arch/sparc/include/asm/pgtable_64.h 	pte = pte_mkdirty(pte);
pte               756 arch/sparc/include/asm/pgtable_64.h 	return __pmd(pte_val(pte));
pte               761 arch/sparc/include/asm/pgtable_64.h 	pte_t pte = __pte(pmd_val(pmd));
pte               763 arch/sparc/include/asm/pgtable_64.h 	pte = pte_mkclean(pte);
pte               765 arch/sparc/include/asm/pgtable_64.h 	return __pmd(pte_val(pte));
pte               770 arch/sparc/include/asm/pgtable_64.h 	pte_t pte = __pte(pmd_val(pmd));
pte               772 arch/sparc/include/asm/pgtable_64.h 	pte = pte_mkyoung(pte);
pte               774 arch/sparc/include/asm/pgtable_64.h 	return __pmd(pte_val(pte));
pte               779 arch/sparc/include/asm/pgtable_64.h 	pte_t pte = __pte(pmd_val(pmd));
pte               781 arch/sparc/include/asm/pgtable_64.h 	pte = pte_mkwrite(pte);
pte               783 arch/sparc/include/asm/pgtable_64.h 	return __pmd(pte_val(pte));
pte               839 arch/sparc/include/asm/pgtable_64.h 	pte_t pte = __pte(pmd_val(pmd));
pte               842 arch/sparc/include/asm/pgtable_64.h 	pfn = pte_pfn(pte);
pte               849 arch/sparc/include/asm/pgtable_64.h 	pte_t pte = __pte(pud_val(pud));
pte               852 arch/sparc/include/asm/pgtable_64.h 	pfn = pte_pfn(pte);
pte               872 arch/sparc/include/asm/pgtable_64.h 	pte_t pte = __pte(pud_val(pud));
pte               874 arch/sparc/include/asm/pgtable_64.h 	return pte_val(pte) & _PAGE_PMD_HUGE;
pte               879 arch/sparc/include/asm/pgtable_64.h 	pte_t pte = __pte(pud_val(pud));
pte               881 arch/sparc/include/asm/pgtable_64.h 	return pte_pfn(pte);
pte               885 arch/sparc/include/asm/pgtable_64.h #define pte_none(pte) 			(!pte_val(pte))
pte               913 arch/sparc/include/asm/pgtable_64.h #define pte_unmap(pte)			do { } while (0)
pte               948 arch/sparc/include/asm/pgtable_64.h 			     pte_t *ptep, pte_t pte, int fullmm)
pte               952 arch/sparc/include/asm/pgtable_64.h 	*ptep = pte;
pte               956 arch/sparc/include/asm/pgtable_64.h #define set_pte_at(mm,addr,ptep,pte)	\
pte               957 arch/sparc/include/asm/pgtable_64.h 	__set_pte_at((mm), (addr), (ptep), (pte), 0)
pte               968 arch/sparc/include/asm/pgtable_64.h #define move_pte(pte, prot, old_addr, new_addr)				\
pte               970 arch/sparc/include/asm/pgtable_64.h 	pte_t newpte = (pte);						\
pte               971 arch/sparc/include/asm/pgtable_64.h 	if (tlb_type != hypervisor && pte_present(pte)) {		\
pte               972 arch/sparc/include/asm/pgtable_64.h 		unsigned long this_pfn = pte_pfn(pte);			\
pte              1018 arch/sparc/include/asm/pgtable_64.h #define __pte_to_swp_entry(pte)		((swp_entry_t) { pte_val(pte) })
pte              1035 arch/sparc/include/asm/pgtable_64.h 		      unsigned long addr, pte_t pte);
pte              1044 arch/sparc/include/asm/pgtable_64.h 				     pte_t pte, pte_t oldpte)
pte              1053 arch/sparc/include/asm/pgtable_64.h 	if (adi_state.enabled && (pte_val(pte) & _PAGE_MCD_4V))
pte              1054 arch/sparc/include/asm/pgtable_64.h 		adi_restore_tags(mm, vma, addr, pte);
pte              1104 arch/sparc/include/asm/pgtable_64.h static inline bool pte_access_permitted(pte_t pte, bool write)
pte              1118 arch/sparc/include/asm/pgtable_64.h 	return (pte_val(pte) & (prot | _PAGE_SPECIAL)) == prot;
pte               328 arch/sparc/kernel/adi_64.c 		      unsigned long addr, pte_t pte)
pte               343 arch/sparc/kernel/adi_64.c 	paddr = pte_val(pte) & _PAGE_PADDR_4V;
pte               300 arch/sparc/kernel/signal32.c 	pte_t *ptep, pte;
pte               329 arch/sparc/kernel/signal32.c 	pte = *ptep;
pte               330 arch/sparc/kernel/signal32.c 	if (!pte_present(pte))
pte               333 arch/sparc/kernel/signal32.c 	paddr = (unsigned long) page_address(pte_page(pte));
pte                85 arch/sparc/mm/fault_64.c 	pte_t *ptep, pte;
pte               115 arch/sparc/mm/fault_64.c 		pte = *ptep;
pte               116 arch/sparc/mm/fault_64.c 		if (pte_present(pte)) {
pte               117 arch/sparc/mm/fault_64.c 			pa  = (pte_pfn(pte) << PAGE_SHIFT);
pte               185 arch/sparc/mm/hugetlbpage.c 	pte_t pte;
pte               187 arch/sparc/mm/hugetlbpage.c 	pte = hugepage_shift_to_tte(entry, shift);
pte               193 arch/sparc/mm/hugetlbpage.c 		return pte_mkmcd(pte);
pte               195 arch/sparc/mm/hugetlbpage.c 		return pte_mknotmcd(pte);
pte               197 arch/sparc/mm/hugetlbpage.c 	return pte;
pte               267 arch/sparc/mm/hugetlbpage.c static unsigned long huge_tte_to_size(pte_t pte)
pte               269 arch/sparc/mm/hugetlbpage.c 	unsigned long size = 1UL << huge_tte_to_shift(pte);
pte               269 arch/sparc/mm/init_64.c static inline void tsb_insert(struct tsb *ent, unsigned long tag, unsigned long pte)
pte               276 arch/sparc/mm/init_64.c 	__tsb_insert(tsb_addr, tag, pte);
pte               421 arch/sparc/mm/init_64.c 	pte_t pte = *ptep;
pte               424 arch/sparc/mm/init_64.c 		unsigned long pfn = pte_pfn(pte);
pte               433 arch/sparc/mm/init_64.c 	if (!pte_accessible(mm, pte))
pte               452 arch/sparc/mm/init_64.c 			pte_val(pte) &= ~mask;
pte               453 arch/sparc/mm/init_64.c 			pte_val(pte) |= (address & mask);
pte               458 arch/sparc/mm/init_64.c 			pte_val(pte) |= (address & (1UL << REAL_HPAGE_SHIFT));
pte               463 arch/sparc/mm/init_64.c 				REAL_HPAGE_SHIFT, address, pte_val(pte));
pte               470 arch/sparc/mm/init_64.c 					address, pte_val(pte));
pte               679 arch/sparc/mm/init_64.c 				       unsigned long pte,
pte               682 arch/sparc/mm/init_64.c 	unsigned long ret = sun4v_mmu_map_perm_addr(vaddr, 0, pte, mmu);
pte               686 arch/sparc/mm/init_64.c 			    "errors with %lx\n", vaddr, 0, pte, mmu, ret);
pte              1658 arch/sparc/mm/init_64.c 	pte_t *pte;
pte              1691 arch/sparc/mm/init_64.c 	pte = pte_offset_kernel(pmd, addr);
pte              1692 arch/sparc/mm/init_64.c 	if (pte_none(*pte))
pte              1695 arch/sparc/mm/init_64.c 	return pfn_valid(pte_pfn(*pte));
pte              1805 arch/sparc/mm/init_64.c 		pte_t *pte;
pte              1849 arch/sparc/mm/init_64.c 		pte = pte_offset_kernel(pmd, vstart);
pte              1855 arch/sparc/mm/init_64.c 			pte_val(*pte) = (paddr | pgprot_val(prot));
pte              1859 arch/sparc/mm/init_64.c 			pte++;
pte              2614 arch/sparc/mm/init_64.c 		unsigned long pte;
pte              2626 arch/sparc/mm/init_64.c 		pte = pmd_val(*pmd);
pte              2627 arch/sparc/mm/init_64.c 		if (!(pte & _PAGE_VALID)) {
pte              2807 arch/sparc/mm/init_64.c 	pte_t pte;
pte              2809 arch/sparc/mm/init_64.c 	pte_val(pte)  = page | pgprot_val(pgprot_noncached(prot));
pte              2810 arch/sparc/mm/init_64.c 	pte_val(pte) |= (((unsigned long)space) << 32);
pte              2811 arch/sparc/mm/init_64.c 	pte_val(pte) |= pte_sz_bits(page_size);
pte              2813 arch/sparc/mm/init_64.c 	return pte;
pte              2893 arch/sparc/mm/init_64.c 	pte_t *pte = NULL;
pte              2896 arch/sparc/mm/init_64.c 		pte = (pte_t *) page_address(page);
pte              2898 arch/sparc/mm/init_64.c 	return pte;
pte              2913 arch/sparc/mm/init_64.c void pte_free_kernel(struct mm_struct *mm, pte_t *pte)
pte              2915 arch/sparc/mm/init_64.c 	free_page((unsigned long)pte);
pte              2918 arch/sparc/mm/init_64.c static void __pte_free(pgtable_t pte)
pte              2920 arch/sparc/mm/init_64.c 	struct page *page = virt_to_page(pte);
pte              2926 arch/sparc/mm/init_64.c void pte_free(struct mm_struct *mm, pgtable_t pte)
pte              2928 arch/sparc/mm/init_64.c 	__pte_free(pte);
pte              2943 arch/sparc/mm/init_64.c 	unsigned long pte, flags;
pte              2950 arch/sparc/mm/init_64.c 	pte = pmd_val(entry);
pte              2953 arch/sparc/mm/init_64.c 	if (!(pte & _PAGE_VALID))
pte              2957 arch/sparc/mm/init_64.c 	pte |= (addr & (1UL << REAL_HPAGE_SHIFT));
pte              2965 arch/sparc/mm/init_64.c 					addr, pte);
pte                42 arch/sparc/mm/leon_mm.c 	unsigned int lvl, pte, paddrbase;
pte                75 arch/sparc/mm/leon_mm.c 		pte = pgd;
pte                98 arch/sparc/mm/leon_mm.c 		pte = pmd;
pte               126 arch/sparc/mm/leon_mm.c 		pte = ped;
pte               149 arch/sparc/mm/leon_mm.c 		pte = ptr;
pte               161 arch/sparc/mm/leon_mm.c 		    (vaddr & ~(-1 << LEON_PTE_SH)) | ((pte & ~0xff) << 4);
pte               165 arch/sparc/mm/leon_mm.c 		    (vaddr & ~(-1 << LEON_PMD_SH)) | ((pte & ~0xff) << 4);
pte               169 arch/sparc/mm/leon_mm.c 		    (vaddr & ~(-1 << LEON_PGD_SH)) | ((pte & ~0xff) << 4);
pte               180 arch/sparc/mm/leon_mm.c 	return pte;
pte               112 arch/sparc/mm/srmmu.c 	pte_t pte;
pte               114 arch/sparc/mm/srmmu.c 	pte = __pte((SRMMU_ET_PTD | (__nocache_pa(pgdp) >> 4)));
pte               115 arch/sparc/mm/srmmu.c 	set_pte((pte_t *)ctxp, pte);
pte               164 arch/sparc/mm/srmmu.c 	void *pte;
pte               166 arch/sparc/mm/srmmu.c 	pte = __nocache_va((dir->pmdv[0] & SRMMU_PTD_PMASK) << 4);
pte               167 arch/sparc/mm/srmmu.c 	return (pte_t *) pte +
pte               300 arch/sparc/mm/srmmu.c 	pte_t *pte;
pte               333 arch/sparc/mm/srmmu.c 		pte = pte_offset_kernel(__nocache_fix(pmd), vaddr);
pte               340 arch/sparc/mm/srmmu.c 		set_pte(__nocache_fix(pte), __pte(pteval));
pte               375 arch/sparc/mm/srmmu.c 	unsigned long pte;
pte               378 arch/sparc/mm/srmmu.c 	if ((pte = (unsigned long)pte_alloc_one_kernel(mm)) == 0)
pte               380 arch/sparc/mm/srmmu.c 	page = pfn_to_page(__nocache_pa(pte) >> PAGE_SHIFT);
pte               388 arch/sparc/mm/srmmu.c void pte_free(struct mm_struct *mm, pgtable_t pte)
pte               392 arch/sparc/mm/srmmu.c 	pgtable_pte_page_dtor(pte);
pte               393 arch/sparc/mm/srmmu.c 	p = (unsigned long)page_address(pte);	/* Cached address (for test) */
pte               396 arch/sparc/mm/srmmu.c 	p = page_to_pfn(pte) << PAGE_SHIFT;	/* Physical address */
pte               910 arch/sparc/mm/srmmu.c 	pte_t *pte;
pte               971 arch/sparc/mm/srmmu.c 	pte = pte_offset_kernel(pmd, PKMAP_BASE);
pte               972 arch/sparc/mm/srmmu.c 	pkmap_page_table = pte;
pte               150 arch/sparc/mm/tlb.c 	pte_t *pte;
pte               152 arch/sparc/mm/tlb.c 	pte = pte_offset_map(&pmd, vaddr);
pte               155 arch/sparc/mm/tlb.c 		if (pte_val(*pte) & _PAGE_VALID) {
pte               156 arch/sparc/mm/tlb.c 			bool exec = pte_exec(*pte);
pte               160 arch/sparc/mm/tlb.c 		pte++;
pte               163 arch/sparc/mm/tlb.c 	pte_unmap(pte);
pte                37 arch/um/include/asm/page.h typedef struct { unsigned long pte; } pte_t;
pte                40 arch/um/include/asm/page.h #define pte_val(p) ((p).pte)
pte                42 arch/um/include/asm/page.h #define pte_get_bits(p, bits) ((p).pte & (bits))
pte                43 arch/um/include/asm/page.h #define pte_set_bits(p, bits) ((p).pte |= (bits))
pte                44 arch/um/include/asm/page.h #define pte_clear_bits(p, bits) ((p).pte &= ~(bits))
pte                45 arch/um/include/asm/page.h #define pte_copy(to, from) ({ (to).pte = (from).pte; })
pte                46 arch/um/include/asm/page.h #define pte_is_zero(p) (!((p).pte & ~_PAGE_NEWPAGE))
pte                48 arch/um/include/asm/page.h 	({ (p).pte = (phys) | pgprot_val(prot); })
pte                57 arch/um/include/asm/page.h typedef struct { unsigned long pte; } pte_t;
pte                66 arch/um/include/asm/page.h #define pte_val(x)	((x).pte)
pte                69 arch/um/include/asm/page.h #define pte_get_bits(p, bits) ((p).pte & (bits))
pte                70 arch/um/include/asm/page.h #define pte_set_bits(p, bits) ((p).pte |= (bits))
pte                71 arch/um/include/asm/page.h #define pte_clear_bits(p, bits) ((p).pte &= ~(bits))
pte                72 arch/um/include/asm/page.h #define pte_copy(to, from) ((to).pte = (from).pte)
pte                73 arch/um/include/asm/page.h #define pte_is_zero(p) (!((p).pte & ~_PAGE_NEWPAGE))
pte                74 arch/um/include/asm/page.h #define pte_set_val(p, phys, prot) (p).pte = (phys | pgprot_val(prot))
pte                15 arch/um/include/asm/pgalloc.h #define pmd_populate_kernel(mm, pmd, pte) \
pte                16 arch/um/include/asm/pgalloc.h 	set_pmd(pmd, __pmd(_PAGE_TABLE + (unsigned long) __pa(pte)))
pte                18 arch/um/include/asm/pgalloc.h #define pmd_populate(mm, pmd, pte) 				\
pte                20 arch/um/include/asm/pgalloc.h 		((unsigned long long)page_to_pfn(pte) <<	\
pte                30 arch/um/include/asm/pgalloc.h #define __pte_free_tlb(tlb,pte, address)		\
pte                32 arch/um/include/asm/pgalloc.h 	pgtable_pte_page_dtor(pte);			\
pte                33 arch/um/include/asm/pgalloc.h 	tlb_remove_page((tlb),(pte));			\
pte                97 arch/um/include/asm/pgtable-3level.h static inline unsigned long pte_pfn(pte_t pte)
pte                99 arch/um/include/asm/pgtable-3level.h 	return phys_to_pfn(pte_val(pte));
pte               104 arch/um/include/asm/pgtable-3level.h 	pte_t pte;
pte               107 arch/um/include/asm/pgtable-3level.h 	pte_set_val(pte, phys, pgprot);
pte               108 arch/um/include/asm/pgtable-3level.h 	return pte;
pte               121 arch/um/include/asm/pgtable.h static inline int pte_none(pte_t pte)
pte               123 arch/um/include/asm/pgtable.h 	return pte_is_zero(pte);
pte               130 arch/um/include/asm/pgtable.h static inline int pte_read(pte_t pte)
pte               132 arch/um/include/asm/pgtable.h 	return((pte_get_bits(pte, _PAGE_USER)) &&
pte               133 arch/um/include/asm/pgtable.h 	       !(pte_get_bits(pte, _PAGE_PROTNONE)));
pte               136 arch/um/include/asm/pgtable.h static inline int pte_exec(pte_t pte){
pte               137 arch/um/include/asm/pgtable.h 	return((pte_get_bits(pte, _PAGE_USER)) &&
pte               138 arch/um/include/asm/pgtable.h 	       !(pte_get_bits(pte, _PAGE_PROTNONE)));
pte               141 arch/um/include/asm/pgtable.h static inline int pte_write(pte_t pte)
pte               143 arch/um/include/asm/pgtable.h 	return((pte_get_bits(pte, _PAGE_RW)) &&
pte               144 arch/um/include/asm/pgtable.h 	       !(pte_get_bits(pte, _PAGE_PROTNONE)));
pte               147 arch/um/include/asm/pgtable.h static inline int pte_dirty(pte_t pte)
pte               149 arch/um/include/asm/pgtable.h 	return pte_get_bits(pte, _PAGE_DIRTY);
pte               152 arch/um/include/asm/pgtable.h static inline int pte_young(pte_t pte)
pte               154 arch/um/include/asm/pgtable.h 	return pte_get_bits(pte, _PAGE_ACCESSED);
pte               157 arch/um/include/asm/pgtable.h static inline int pte_newpage(pte_t pte)
pte               159 arch/um/include/asm/pgtable.h 	return pte_get_bits(pte, _PAGE_NEWPAGE);
pte               162 arch/um/include/asm/pgtable.h static inline int pte_newprot(pte_t pte)
pte               164 arch/um/include/asm/pgtable.h 	return(pte_present(pte) && (pte_get_bits(pte, _PAGE_NEWPROT)));
pte               167 arch/um/include/asm/pgtable.h static inline int pte_special(pte_t pte)
pte               178 arch/um/include/asm/pgtable.h static inline pte_t pte_mknewprot(pte_t pte)
pte               180 arch/um/include/asm/pgtable.h 	pte_set_bits(pte, _PAGE_NEWPROT);
pte               181 arch/um/include/asm/pgtable.h 	return(pte);
pte               184 arch/um/include/asm/pgtable.h static inline pte_t pte_mkclean(pte_t pte)
pte               186 arch/um/include/asm/pgtable.h 	pte_clear_bits(pte, _PAGE_DIRTY);
pte               187 arch/um/include/asm/pgtable.h 	return(pte);
pte               190 arch/um/include/asm/pgtable.h static inline pte_t pte_mkold(pte_t pte)	
pte               192 arch/um/include/asm/pgtable.h 	pte_clear_bits(pte, _PAGE_ACCESSED);
pte               193 arch/um/include/asm/pgtable.h 	return(pte);
pte               196 arch/um/include/asm/pgtable.h static inline pte_t pte_wrprotect(pte_t pte)
pte               198 arch/um/include/asm/pgtable.h 	if (likely(pte_get_bits(pte, _PAGE_RW)))
pte               199 arch/um/include/asm/pgtable.h 		pte_clear_bits(pte, _PAGE_RW);
pte               201 arch/um/include/asm/pgtable.h 		return pte;
pte               202 arch/um/include/asm/pgtable.h 	return(pte_mknewprot(pte)); 
pte               205 arch/um/include/asm/pgtable.h static inline pte_t pte_mkread(pte_t pte)
pte               207 arch/um/include/asm/pgtable.h 	if (unlikely(pte_get_bits(pte, _PAGE_USER)))
pte               208 arch/um/include/asm/pgtable.h 		return pte;
pte               209 arch/um/include/asm/pgtable.h 	pte_set_bits(pte, _PAGE_USER);
pte               210 arch/um/include/asm/pgtable.h 	return(pte_mknewprot(pte)); 
pte               213 arch/um/include/asm/pgtable.h static inline pte_t pte_mkdirty(pte_t pte)
pte               215 arch/um/include/asm/pgtable.h 	pte_set_bits(pte, _PAGE_DIRTY);
pte               216 arch/um/include/asm/pgtable.h 	return(pte);
pte               219 arch/um/include/asm/pgtable.h static inline pte_t pte_mkyoung(pte_t pte)
pte               221 arch/um/include/asm/pgtable.h 	pte_set_bits(pte, _PAGE_ACCESSED);
pte               222 arch/um/include/asm/pgtable.h 	return(pte);
pte               225 arch/um/include/asm/pgtable.h static inline pte_t pte_mkwrite(pte_t pte)	
pte               227 arch/um/include/asm/pgtable.h 	if (unlikely(pte_get_bits(pte,  _PAGE_RW)))
pte               228 arch/um/include/asm/pgtable.h 		return pte;
pte               229 arch/um/include/asm/pgtable.h 	pte_set_bits(pte, _PAGE_RW);
pte               230 arch/um/include/asm/pgtable.h 	return(pte_mknewprot(pte)); 
pte               233 arch/um/include/asm/pgtable.h static inline pte_t pte_mkuptodate(pte_t pte)	
pte               235 arch/um/include/asm/pgtable.h 	pte_clear_bits(pte, _PAGE_NEWPAGE);
pte               236 arch/um/include/asm/pgtable.h 	if(pte_present(pte))
pte               237 arch/um/include/asm/pgtable.h 		pte_clear_bits(pte, _PAGE_NEWPROT);
pte               238 arch/um/include/asm/pgtable.h 	return(pte); 
pte               241 arch/um/include/asm/pgtable.h static inline pte_t pte_mknewpage(pte_t pte)
pte               243 arch/um/include/asm/pgtable.h 	pte_set_bits(pte, _PAGE_NEWPAGE);
pte               244 arch/um/include/asm/pgtable.h 	return(pte);
pte               247 arch/um/include/asm/pgtable.h static inline pte_t pte_mkspecial(pte_t pte)
pte               249 arch/um/include/asm/pgtable.h 	return(pte);
pte               288 arch/um/include/asm/pgtable.h 	({ pte_t pte;					\
pte               290 arch/um/include/asm/pgtable.h 	pte_set_val(pte, page_to_phys(page), (pgprot));	\
pte               291 arch/um/include/asm/pgtable.h 	if (pte_present(pte))				\
pte               292 arch/um/include/asm/pgtable.h 		pte_mknewprot(pte_mknewpage(pte));	\
pte               293 arch/um/include/asm/pgtable.h 	pte;})
pte               295 arch/um/include/asm/pgtable.h static inline pte_t pte_modify(pte_t pte, pgprot_t newprot)
pte               297 arch/um/include/asm/pgtable.h 	pte_set_val(pte, (pte_val(pte) & _PAGE_CHG_MASK), newprot);
pte               298 arch/um/include/asm/pgtable.h 	return pte; 
pte               344 arch/um/include/asm/pgtable.h #define pte_unmap(pte) do { } while (0)
pte               357 arch/um/include/asm/pgtable.h #define __pte_to_swp_entry(pte) \
pte               358 arch/um/include/asm/pgtable.h 	((swp_entry_t) { pte_val(pte_mkuptodate(pte)) })
pte                68 arch/um/kernel/mem.c 		pte_t *pte = (pte_t *) memblock_alloc_low(PAGE_SIZE,
pte                70 arch/um/kernel/mem.c 		if (!pte)
pte                75 arch/um/kernel/mem.c 					   (unsigned long) __pa(pte)));
pte                76 arch/um/kernel/mem.c 		if (pte != pte_offset_kernel(pmd, 0))
pte               129 arch/um/kernel/mem.c 	pte_t *pte;
pte               149 arch/um/kernel/mem.c 		pte = pte_offset_kernel(pmd, vaddr);
pte               150 arch/um/kernel/mem.c 		pte_set_val(*pte, p, PAGE_READONLY);
pte                24 arch/um/kernel/skas/mmu.c 	pte_t *pte;
pte                35 arch/um/kernel/skas/mmu.c 	pte = pte_alloc_map(mm, pmd, proc);
pte                36 arch/um/kernel/skas/mmu.c 	if (!pte)
pte                39 arch/um/kernel/skas/mmu.c 	*pte = mk_pte(virt_to_page(kernel), __pgprot(_PAGE_PRESENT));
pte                40 arch/um/kernel/skas/mmu.c 	*pte = pte_mkread(*pte);
pte               127 arch/um/kernel/skas/mmu.c 	pte_t *pte;
pte               129 arch/um/kernel/skas/mmu.c 	pte = virt_to_pte(mm, STUB_CODE);
pte               130 arch/um/kernel/skas/mmu.c 	if (pte != NULL)
pte               131 arch/um/kernel/skas/mmu.c 		pte_clear(mm, STUB_CODE, pte);
pte               133 arch/um/kernel/skas/mmu.c 	pte = virt_to_pte(mm, STUB_DATA);
pte               134 arch/um/kernel/skas/mmu.c 	if (pte == NULL)
pte               137 arch/um/kernel/skas/mmu.c 	pte_clear(mm, STUB_DATA, pte);
pte                43 arch/um/kernel/skas/uaccess.c 	pte_t *pte = virt_to_pte(current->mm, virt);
pte                46 arch/um/kernel/skas/uaccess.c 	if ((pte == NULL) || !pte_present(*pte) ||
pte                47 arch/um/kernel/skas/uaccess.c 	    (is_write && !pte_write(*pte))) {
pte                51 arch/um/kernel/skas/uaccess.c 		pte = virt_to_pte(current->mm, virt);
pte                53 arch/um/kernel/skas/uaccess.c 	if (!pte_present(*pte))
pte                54 arch/um/kernel/skas/uaccess.c 		pte = NULL;
pte                56 arch/um/kernel/skas/uaccess.c 	return pte;
pte                63 arch/um/kernel/skas/uaccess.c 	pte_t *pte;
pte                66 arch/um/kernel/skas/uaccess.c 	pte = maybe_map(addr, is_write);
pte                67 arch/um/kernel/skas/uaccess.c 	if (pte == NULL)
pte                70 arch/um/kernel/skas/uaccess.c 	page = pte_page(*pte);
pte               225 arch/um/kernel/tlb.c 	pte_t *pte;
pte               228 arch/um/kernel/tlb.c 	pte = pte_offset_kernel(pmd, addr);
pte               233 arch/um/kernel/tlb.c 		r = pte_read(*pte);
pte               234 arch/um/kernel/tlb.c 		w = pte_write(*pte);
pte               235 arch/um/kernel/tlb.c 		x = pte_exec(*pte);
pte               236 arch/um/kernel/tlb.c 		if (!pte_young(*pte)) {
pte               239 arch/um/kernel/tlb.c 		} else if (!pte_dirty(*pte))
pte               244 arch/um/kernel/tlb.c 		if (hvc->force || pte_newpage(*pte)) {
pte               245 arch/um/kernel/tlb.c 			if (pte_present(*pte)) {
pte               246 arch/um/kernel/tlb.c 				if (pte_newpage(*pte))
pte               247 arch/um/kernel/tlb.c 					ret = add_mmap(addr, pte_val(*pte) & PAGE_MASK,
pte               251 arch/um/kernel/tlb.c 		} else if (pte_newprot(*pte))
pte               253 arch/um/kernel/tlb.c 		*pte = pte_mkuptodate(*pte);
pte               254 arch/um/kernel/tlb.c 	} while (pte++, addr += PAGE_SIZE, ((addr < end) && !ret));
pte               343 arch/um/kernel/tlb.c 	pte_t *pte;
pte               399 arch/um/kernel/tlb.c 		pte = pte_offset_kernel(pmd, addr);
pte               400 arch/um/kernel/tlb.c 		if (!pte_present(*pte) || pte_newpage(*pte)) {
pte               406 arch/um/kernel/tlb.c 			if (pte_present(*pte))
pte               407 arch/um/kernel/tlb.c 				err = add_mmap(addr, pte_val(*pte) & PAGE_MASK,
pte               410 arch/um/kernel/tlb.c 		else if (pte_newprot(*pte)) {
pte               429 arch/um/kernel/tlb.c 	pte_t *pte;
pte               448 arch/um/kernel/tlb.c 	pte = pte_offset_kernel(pmd, address);
pte               450 arch/um/kernel/tlb.c 	r = pte_read(*pte);
pte               451 arch/um/kernel/tlb.c 	w = pte_write(*pte);
pte               452 arch/um/kernel/tlb.c 	x = pte_exec(*pte);
pte               453 arch/um/kernel/tlb.c 	if (!pte_young(*pte)) {
pte               456 arch/um/kernel/tlb.c 	} else if (!pte_dirty(*pte)) {
pte               463 arch/um/kernel/tlb.c 	if (pte_newpage(*pte)) {
pte               464 arch/um/kernel/tlb.c 		if (pte_present(*pte)) {
pte               468 arch/um/kernel/tlb.c 			fd = phys_mapping(pte_val(*pte) & PAGE_MASK, &offset);
pte               474 arch/um/kernel/tlb.c 	else if (pte_newprot(*pte))
pte               484 arch/um/kernel/tlb.c 	*pte = pte_mkuptodate(*pte);
pte                33 arch/um/kernel/trap.c 	pte_t *pte;
pte               109 arch/um/kernel/trap.c 		pte = pte_offset_kernel(pmd, address);
pte               110 arch/um/kernel/trap.c 	} while (!pte_present(*pte));
pte               121 arch/um/kernel/trap.c 	WARN_ON(!pte_young(*pte) || (is_write && !pte_dirty(*pte)));
pte                36 arch/unicore32/include/asm/cpu-single.h extern void cpu_set_pte(pte_t *ptep, pte_t pte);
pte                34 arch/unicore32/include/asm/page.h typedef struct { unsigned long pte; } pte_t;
pte                38 arch/unicore32/include/asm/page.h #define pte_val(x)      ((x).pte)
pte                36 arch/unicore32/include/asm/pgalloc.h 	pte_t *pte = __pte_alloc_one_kernel(mm);
pte                38 arch/unicore32/include/asm/pgalloc.h 	if (pte)
pte                39 arch/unicore32/include/asm/pgalloc.h 		clean_dcache_area(pte, PTRS_PER_PTE * sizeof(pte_t));
pte                41 arch/unicore32/include/asm/pgalloc.h 	return pte;
pte                47 arch/unicore32/include/asm/pgalloc.h 	struct page *pte;
pte                49 arch/unicore32/include/asm/pgalloc.h 	pte = __pte_alloc_one(mm, GFP_PGTABLE_USER);
pte                50 arch/unicore32/include/asm/pgalloc.h 	if (!pte)
pte                52 arch/unicore32/include/asm/pgalloc.h 	if (!PageHighMem(pte))
pte                53 arch/unicore32/include/asm/pgalloc.h 		clean_pte_table(page_address(pte));
pte                54 arch/unicore32/include/asm/pgalloc.h 	return pte;
pte                50 arch/unicore32/include/asm/pgtable.h #define pte_ERROR(pte)		__pte_error(__FILE__, __LINE__, pte_val(pte))
pte               150 arch/unicore32/include/asm/pgtable.h #define pte_pfn(pte)			(pte_val(pte) >> PAGE_SHIFT)
pte               154 arch/unicore32/include/asm/pgtable.h #define pte_none(pte)			(!pte_val(pte))
pte               156 arch/unicore32/include/asm/pgtable.h #define pte_page(pte)			(pfn_to_page(pte_pfn(pte)))
pte               162 arch/unicore32/include/asm/pgtable.h #define pte_unmap(pte)			do { } while (0)
pte               164 arch/unicore32/include/asm/pgtable.h #define set_pte(ptep, pte)	cpu_set_pte(ptep, pte)
pte               175 arch/unicore32/include/asm/pgtable.h #define pte_present(pte)	(pte_val(pte) & PTE_PRESENT)
pte               176 arch/unicore32/include/asm/pgtable.h #define pte_write(pte)		(pte_val(pte) & PTE_WRITE)
pte               177 arch/unicore32/include/asm/pgtable.h #define pte_dirty(pte)		(pte_val(pte) & PTE_DIRTY)
pte               178 arch/unicore32/include/asm/pgtable.h #define pte_young(pte)		(pte_val(pte) & PTE_YOUNG)
pte               179 arch/unicore32/include/asm/pgtable.h #define pte_exec(pte)		(pte_val(pte) & PTE_EXEC)
pte               180 arch/unicore32/include/asm/pgtable.h #define pte_special(pte)	(0)
pte               183 arch/unicore32/include/asm/pgtable.h static inline pte_t pte_##fn(pte_t pte) { pte_val(pte) op; return pte; }
pte               192 arch/unicore32/include/asm/pgtable.h static inline pte_t pte_mkspecial(pte_t pte) { return pte; }
pte               239 arch/unicore32/include/asm/pgtable.h static inline pte_t pte_modify(pte_t pte, pgprot_t newprot)
pte               242 arch/unicore32/include/asm/pgtable.h 	pte_val(pte) = (pte_val(pte) & ~mask) | (pgprot_val(newprot) & mask);
pte               243 arch/unicore32/include/asm/pgtable.h 	return pte;
pte               271 arch/unicore32/include/asm/pgtable.h #define __pte_to_swp_entry(pte)	((swp_entry_t) { pte_val(pte) })
pte                16 arch/unicore32/include/asm/tlb.h #define __pte_free_tlb(tlb, pte, addr)				\
pte                18 arch/unicore32/include/asm/tlb.h 		pgtable_pte_page_dtor(pte);			\
pte                19 arch/unicore32/include/asm/tlb.h 		tlb_remove_page((tlb), (pte));			\
pte                75 arch/unicore32/kernel/hibernate.c 	pte_t *pte;
pte                99 arch/unicore32/kernel/hibernate.c 			pte = resume_one_page_table_init(pmd);
pte               100 arch/unicore32/kernel/hibernate.c 			if (!pte)
pte               103 arch/unicore32/kernel/hibernate.c 			max_pte = pte + PTRS_PER_PTE;
pte               104 arch/unicore32/kernel/hibernate.c 			for (; pte < max_pte; pte++, pfn++) {
pte               108 arch/unicore32/kernel/hibernate.c 				set_pte(pte, pfn_pte(pfn, PAGE_KERNEL_EXEC));
pte                51 arch/unicore32/mm/fault.c 		pte_t *pte;
pte                77 arch/unicore32/mm/fault.c 		pte = pte_offset_map(pmd, addr);
pte                78 arch/unicore32/mm/fault.c 		printk(", *pte=%08lx", pte_val(*pte));
pte                79 arch/unicore32/mm/fault.c 		pte_unmap(pte);
pte               146 arch/unicore32/mm/mmu.c 		pte_t *pte = memblock_alloc(size, size);
pte               148 arch/unicore32/mm/mmu.c 		if (!pte)
pte               152 arch/unicore32/mm/mmu.c 		__pmd_populate(pmd, __pa(pte) | prot);
pte               162 arch/unicore32/mm/mmu.c 	pte_t *pte = early_pte_alloc(pmd, addr, type->prot_l1);
pte               164 arch/unicore32/mm/mmu.c 		set_pte(pte, pfn_pte(pfn, __pgprot(type->prot_pte)));
pte               166 arch/unicore32/mm/mmu.c 	} while (pte++, addr += PAGE_SIZE, addr != end);
pte                79 arch/unicore32/mm/pgd.c 	pgtable_t pte;
pte                94 arch/unicore32/mm/pgd.c 	pte = pmd_pgtable(*pmd);
pte                96 arch/unicore32/mm/pgd.c 	pte_free(mm, pte);
pte               158 arch/x86/include/asm/fixmap.h void __native_set_fixmap(enum fixed_addresses idx, pte_t pte);
pte               156 arch/x86/include/asm/kexec.h 	pte_t *pte;
pte               395 arch/x86/include/asm/kvm_host.h 			   u64 *spte, const void *pte);
pte              1559 arch/x86/include/asm/kvm_host.h int kvm_set_spte_hva(struct kvm *kvm, unsigned long hva, pte_t pte);
pte               371 arch/x86/include/asm/paravirt.h 	return (pte_t) { .pte = ret };
pte               374 arch/x86/include/asm/paravirt.h static inline pteval_t pte_val(pte_t pte)
pte               380 arch/x86/include/asm/paravirt.h 				   pte.pte, (u64)pte.pte >> 32);
pte               382 arch/x86/include/asm/paravirt.h 		ret = PVOP_CALLEE1(pteval_t, mmu.pte_val, pte.pte);
pte               420 arch/x86/include/asm/paravirt.h 	return (pte_t) { .pte = ret };
pte               424 arch/x86/include/asm/paravirt.h 					   pte_t *ptep, pte_t old_pte, pte_t pte)
pte               429 arch/x86/include/asm/paravirt.h 		pv_ops.mmu.ptep_modify_prot_commit(vma, addr, ptep, pte);
pte               432 arch/x86/include/asm/paravirt.h 			    vma, addr, ptep, pte.pte);
pte               435 arch/x86/include/asm/paravirt.h static inline void set_pte(pte_t *ptep, pte_t pte)
pte               438 arch/x86/include/asm/paravirt.h 		PVOP_VCALL3(mmu.set_pte, ptep, pte.pte, (u64)pte.pte >> 32);
pte               440 arch/x86/include/asm/paravirt.h 		PVOP_VCALL2(mmu.set_pte, ptep, pte.pte);
pte               444 arch/x86/include/asm/paravirt.h 			      pte_t *ptep, pte_t pte)
pte               448 arch/x86/include/asm/paravirt.h 		pv_ops.mmu.set_pte_at(mm, addr, ptep, pte);
pte               450 arch/x86/include/asm/paravirt.h 		PVOP_VCALL4(mmu.set_pte_at, mm, addr, ptep, pte.pte);
pte               570 arch/x86/include/asm/paravirt.h static inline void set_pte_atomic(pte_t *ptep, pte_t pte)
pte               572 arch/x86/include/asm/paravirt.h 	PVOP_VCALL3(mmu.set_pte_atomic, ptep, pte.pte, pte.pte >> 32);
pte               586 arch/x86/include/asm/paravirt.h static inline void set_pte_atomic(pte_t *ptep, pte_t pte)
pte               588 arch/x86/include/asm/paravirt.h 	set_pte(ptep, pte);
pte               256 arch/x86/include/asm/paravirt_types.h 					pte_t *ptep, pte_t pte);
pte                55 arch/x86/include/asm/pgalloc.h extern void ___pte_free_tlb(struct mmu_gather *tlb, struct page *pte);
pte                57 arch/x86/include/asm/pgalloc.h static inline void __pte_free_tlb(struct mmu_gather *tlb, struct page *pte,
pte                60 arch/x86/include/asm/pgalloc.h 	___pte_free_tlb(tlb, pte);
pte                64 arch/x86/include/asm/pgalloc.h 				       pmd_t *pmd, pte_t *pte)
pte                66 arch/x86/include/asm/pgalloc.h 	paravirt_alloc_pte(mm, __pa(pte) >> PAGE_SHIFT);
pte                67 arch/x86/include/asm/pgalloc.h 	set_pmd(pmd, __pmd(__pa(pte) | _PAGE_TABLE));
pte                71 arch/x86/include/asm/pgalloc.h 				       pmd_t *pmd, pte_t *pte)
pte                73 arch/x86/include/asm/pgalloc.h 	paravirt_alloc_pte(mm, __pa(pte) >> PAGE_SHIFT);
pte                74 arch/x86/include/asm/pgalloc.h 	set_pmd_safe(pmd, __pmd(__pa(pte) | _PAGE_TABLE));
pte                78 arch/x86/include/asm/pgalloc.h 				struct page *pte)
pte                80 arch/x86/include/asm/pgalloc.h 	unsigned long pfn = page_to_pfn(pte);
pte                15 arch/x86/include/asm/pgtable-2level.h static inline void native_set_pte(pte_t *ptep , pte_t pte)
pte                17 arch/x86/include/asm/pgtable-2level.h 	*ptep = pte;
pte                29 arch/x86/include/asm/pgtable-2level.h static inline void native_set_pte_atomic(pte_t *ptep, pte_t pte)
pte                31 arch/x86/include/asm/pgtable-2level.h 	native_set_pte(ptep, pte);
pte                95 arch/x86/include/asm/pgtable-2level.h #define __pte_to_swp_entry(pte)		((swp_entry_t) { (pte).pte_low })
pte                96 arch/x86/include/asm/pgtable-2level.h #define __swp_entry_to_pte(x)		((pte_t) { .pte = (x).val })
pte                16 arch/x86/include/asm/pgtable-2level_types.h 	pteval_t pte;
pte                30 arch/x86/include/asm/pgtable-3level.h static inline void native_set_pte(pte_t *ptep, pte_t pte)
pte                32 arch/x86/include/asm/pgtable-3level.h 	ptep->pte_high = pte.pte_high;
pte                34 arch/x86/include/asm/pgtable-3level.h 	ptep->pte_low = pte.pte_low;
pte                91 arch/x86/include/asm/pgtable-3level.h static inline void native_set_pte_atomic(pte_t *ptep, pte_t pte)
pte                93 arch/x86/include/asm/pgtable-3level.h 	set_64bit((unsigned long long *)(ptep), native_pte_val(pte));
pte               155 arch/x86/include/asm/pgtable-3level.h 	res.pte = (pteval_t)arch_atomic64_xchg((atomic64_t *)ptep, 0);
pte               274 arch/x86/include/asm/pgtable-3level.h #define __swp_entry_to_pte(x)	((pte_t){ .pte = \
pte               282 arch/x86/include/asm/pgtable-3level.h #define __pteval_swp_type(x) ((unsigned long)((x).pte >> (64 - SWP_TYPE_BITS)))
pte               283 arch/x86/include/asm/pgtable-3level.h #define __pteval_swp_offset(x) ((unsigned long)(~((x).pte) << SWP_TYPE_BITS >> SWP_OFFSET_SHIFT))
pte               285 arch/x86/include/asm/pgtable-3level.h #define __pte_to_swp_entry(pte)	(__swp_entry(__pteval_swp_type(pte), \
pte               286 arch/x86/include/asm/pgtable-3level.h 					     __pteval_swp_offset(pte)))
pte                19 arch/x86/include/asm/pgtable-3level_types.h 	pteval_t pte;
pte                63 arch/x86/include/asm/pgtable.h #define set_pte(ptep, pte)		native_set_pte(ptep, pte)
pte                64 arch/x86/include/asm/pgtable.h #define set_pte_at(mm, addr, ptep, pte)	native_set_pte_at(mm, addr, ptep, pte)
pte                66 arch/x86/include/asm/pgtable.h #define set_pte_atomic(ptep, pte)					\
pte                67 arch/x86/include/asm/pgtable.h 	native_set_pte_atomic(ptep, pte)
pte               123 arch/x86/include/asm/pgtable.h static inline int pte_dirty(pte_t pte)
pte               125 arch/x86/include/asm/pgtable.h 	return pte_flags(pte) & _PAGE_DIRTY;
pte               157 arch/x86/include/asm/pgtable.h static inline int pte_young(pte_t pte)
pte               159 arch/x86/include/asm/pgtable.h 	return pte_flags(pte) & _PAGE_ACCESSED;
pte               182 arch/x86/include/asm/pgtable.h static inline int pte_write(pte_t pte)
pte               184 arch/x86/include/asm/pgtable.h 	return pte_flags(pte) & _PAGE_RW;
pte               187 arch/x86/include/asm/pgtable.h static inline int pte_huge(pte_t pte)
pte               189 arch/x86/include/asm/pgtable.h 	return pte_flags(pte) & _PAGE_PSE;
pte               192 arch/x86/include/asm/pgtable.h static inline int pte_global(pte_t pte)
pte               194 arch/x86/include/asm/pgtable.h 	return pte_flags(pte) & _PAGE_GLOBAL;
pte               197 arch/x86/include/asm/pgtable.h static inline int pte_exec(pte_t pte)
pte               199 arch/x86/include/asm/pgtable.h 	return !(pte_flags(pte) & _PAGE_NX);
pte               202 arch/x86/include/asm/pgtable.h static inline int pte_special(pte_t pte)
pte               204 arch/x86/include/asm/pgtable.h 	return pte_flags(pte) & _PAGE_SPECIAL;
pte               211 arch/x86/include/asm/pgtable.h static inline unsigned long pte_pfn(pte_t pte)
pte               213 arch/x86/include/asm/pgtable.h 	phys_addr_t pfn = pte_val(pte);
pte               248 arch/x86/include/asm/pgtable.h #define pte_page(pte)	pfn_to_page(pte_pfn(pte))
pte               250 arch/x86/include/asm/pgtable.h static inline int pmd_large(pmd_t pte)
pte               252 arch/x86/include/asm/pgtable.h 	return pmd_flags(pte) & _PAGE_PSE;
pte               300 arch/x86/include/asm/pgtable.h static inline pte_t pte_set_flags(pte_t pte, pteval_t set)
pte               302 arch/x86/include/asm/pgtable.h 	pteval_t v = native_pte_val(pte);
pte               307 arch/x86/include/asm/pgtable.h static inline pte_t pte_clear_flags(pte_t pte, pteval_t clear)
pte               309 arch/x86/include/asm/pgtable.h 	pteval_t v = native_pte_val(pte);
pte               314 arch/x86/include/asm/pgtable.h static inline pte_t pte_mkclean(pte_t pte)
pte               316 arch/x86/include/asm/pgtable.h 	return pte_clear_flags(pte, _PAGE_DIRTY);
pte               319 arch/x86/include/asm/pgtable.h static inline pte_t pte_mkold(pte_t pte)
pte               321 arch/x86/include/asm/pgtable.h 	return pte_clear_flags(pte, _PAGE_ACCESSED);
pte               324 arch/x86/include/asm/pgtable.h static inline pte_t pte_wrprotect(pte_t pte)
pte               326 arch/x86/include/asm/pgtable.h 	return pte_clear_flags(pte, _PAGE_RW);
pte               329 arch/x86/include/asm/pgtable.h static inline pte_t pte_mkexec(pte_t pte)
pte               331 arch/x86/include/asm/pgtable.h 	return pte_clear_flags(pte, _PAGE_NX);
pte               334 arch/x86/include/asm/pgtable.h static inline pte_t pte_mkdirty(pte_t pte)
pte               336 arch/x86/include/asm/pgtable.h 	return pte_set_flags(pte, _PAGE_DIRTY | _PAGE_SOFT_DIRTY);
pte               339 arch/x86/include/asm/pgtable.h static inline pte_t pte_mkyoung(pte_t pte)
pte               341 arch/x86/include/asm/pgtable.h 	return pte_set_flags(pte, _PAGE_ACCESSED);
pte               344 arch/x86/include/asm/pgtable.h static inline pte_t pte_mkwrite(pte_t pte)
pte               346 arch/x86/include/asm/pgtable.h 	return pte_set_flags(pte, _PAGE_RW);
pte               349 arch/x86/include/asm/pgtable.h static inline pte_t pte_mkhuge(pte_t pte)
pte               351 arch/x86/include/asm/pgtable.h 	return pte_set_flags(pte, _PAGE_PSE);
pte               354 arch/x86/include/asm/pgtable.h static inline pte_t pte_clrhuge(pte_t pte)
pte               356 arch/x86/include/asm/pgtable.h 	return pte_clear_flags(pte, _PAGE_PSE);
pte               359 arch/x86/include/asm/pgtable.h static inline pte_t pte_mkglobal(pte_t pte)
pte               361 arch/x86/include/asm/pgtable.h 	return pte_set_flags(pte, _PAGE_GLOBAL);
pte               364 arch/x86/include/asm/pgtable.h static inline pte_t pte_clrglobal(pte_t pte)
pte               366 arch/x86/include/asm/pgtable.h 	return pte_clear_flags(pte, _PAGE_GLOBAL);
pte               369 arch/x86/include/asm/pgtable.h static inline pte_t pte_mkspecial(pte_t pte)
pte               371 arch/x86/include/asm/pgtable.h 	return pte_set_flags(pte, _PAGE_SPECIAL);
pte               374 arch/x86/include/asm/pgtable.h static inline pte_t pte_mkdevmap(pte_t pte)
pte               376 arch/x86/include/asm/pgtable.h 	return pte_set_flags(pte, _PAGE_SPECIAL|_PAGE_DEVMAP);
pte               488 arch/x86/include/asm/pgtable.h static inline int pte_soft_dirty(pte_t pte)
pte               490 arch/x86/include/asm/pgtable.h 	return pte_flags(pte) & _PAGE_SOFT_DIRTY;
pte               503 arch/x86/include/asm/pgtable.h static inline pte_t pte_mksoft_dirty(pte_t pte)
pte               505 arch/x86/include/asm/pgtable.h 	return pte_set_flags(pte, _PAGE_SOFT_DIRTY);
pte               518 arch/x86/include/asm/pgtable.h static inline pte_t pte_clear_soft_dirty(pte_t pte)
pte               520 arch/x86/include/asm/pgtable.h 	return pte_clear_flags(pte, _PAGE_SOFT_DIRTY);
pte               604 arch/x86/include/asm/pgtable.h static inline pte_t pte_modify(pte_t pte, pgprot_t newprot)
pte               606 arch/x86/include/asm/pgtable.h 	pteval_t val = pte_val(pte), oldval = val;
pte               723 arch/x86/include/asm/pgtable.h static inline int pte_none(pte_t pte)
pte               725 arch/x86/include/asm/pgtable.h 	return !(pte.pte & ~(_PAGE_KNL_ERRATUM_MASK));
pte               731 arch/x86/include/asm/pgtable.h 	return a.pte == b.pte;
pte               775 arch/x86/include/asm/pgtable.h static inline int pte_protnone(pte_t pte)
pte               777 arch/x86/include/asm/pgtable.h 	return (pte_flags(pte) & (_PAGE_PROTNONE | _PAGE_PRESENT))
pte              1084 arch/x86/include/asm/pgtable.h 				     pte_t *ptep , pte_t pte)
pte              1086 arch/x86/include/asm/pgtable.h 	native_set_pte(ptep, pte);
pte              1127 arch/x86/include/asm/pgtable.h 	pte_t pte = native_ptep_get_and_clear(ptep);
pte              1128 arch/x86/include/asm/pgtable.h 	return pte;
pte              1136 arch/x86/include/asm/pgtable.h 	pte_t pte;
pte              1142 arch/x86/include/asm/pgtable.h 		pte = native_local_ptep_get_and_clear(ptep);
pte              1144 arch/x86/include/asm/pgtable.h 		pte = ptep_get_and_clear(mm, addr, ptep);
pte              1146 arch/x86/include/asm/pgtable.h 	return pte;
pte              1153 arch/x86/include/asm/pgtable.h 	clear_bit(_PAGE_BIT_RW, (unsigned long *)&ptep->pte);
pte              1345 arch/x86/include/asm/pgtable.h static inline pte_t pte_swp_mksoft_dirty(pte_t pte)
pte              1347 arch/x86/include/asm/pgtable.h 	return pte_set_flags(pte, _PAGE_SWP_SOFT_DIRTY);
pte              1350 arch/x86/include/asm/pgtable.h static inline int pte_swp_soft_dirty(pte_t pte)
pte              1352 arch/x86/include/asm/pgtable.h 	return pte_flags(pte) & _PAGE_SWP_SOFT_DIRTY;
pte              1355 arch/x86/include/asm/pgtable.h static inline pte_t pte_swp_clear_soft_dirty(pte_t pte)
pte              1357 arch/x86/include/asm/pgtable.h 	return pte_clear_flags(pte, _PAGE_SWP_SOFT_DIRTY);
pte              1445 arch/x86/include/asm/pgtable.h static inline bool pte_access_permitted(pte_t pte, bool write)
pte              1447 arch/x86/include/asm/pgtable.h 	return __pte_access_permitted(pte_val(pte), write);
pte                52 arch/x86/include/asm/pgtable_32.h #define pte_unmap(pte) kunmap_atomic((pte))
pte                56 arch/x86/include/asm/pgtable_32.h #define pte_unmap(pte) do { } while (0)
pte                59 arch/x86/include/asm/pgtable_64.h static inline void native_set_pte(pte_t *ptep, pte_t pte)
pte                61 arch/x86/include/asm/pgtable_64.h 	WRITE_ONCE(*ptep, pte);
pte                70 arch/x86/include/asm/pgtable_64.h static inline void native_set_pte_atomic(pte_t *ptep, pte_t pte)
pte                72 arch/x86/include/asm/pgtable_64.h 	native_set_pte(ptep, pte);
pte                88 arch/x86/include/asm/pgtable_64.h 	return native_make_pte(xchg(&xp->pte, 0));
pte               185 arch/x86/include/asm/pgtable_64.h #define pte_unmap(pte) ((void)(pte))/* NOP */
pte               233 arch/x86/include/asm/pgtable_64.h #define __pte_to_swp_entry(pte)		((swp_entry_t) { pte_val((pte)) })
pte               235 arch/x86/include/asm/pgtable_64.h #define __swp_entry_to_pte(x)		((pte_t) { .pte = (x).val })
pte                21 arch/x86/include/asm/pgtable_64_types.h typedef struct { pteval_t pte; } pte_t;
pte               439 arch/x86/include/asm/pgtable_types.h 	return (pte_t) { .pte = val };
pte               442 arch/x86/include/asm/pgtable_types.h static inline pteval_t native_pte_val(pte_t pte)
pte               444 arch/x86/include/asm/pgtable_types.h 	return pte.pte;
pte               447 arch/x86/include/asm/pgtable_types.h static inline pteval_t pte_flags(pte_t pte)
pte               449 arch/x86/include/asm/pgtable_types.h 	return native_pte_val(pte) & PTE_FLAGS_MASK;
pte               529 arch/x86/include/asm/pgtable_types.h void set_pte_vaddr(unsigned long vaddr, pte_t pte);
pte               349 arch/x86/include/asm/xen/hypercall.h 				   new_val.pte, flags);
pte               352 arch/x86/include/asm/xen/hypercall.h 				   new_val.pte, new_val.pte >> 32, flags);
pte               465 arch/x86/include/asm/xen/hypercall.h 		mcl->args[1] = new_val.pte;
pte               468 arch/x86/include/asm/xen/hypercall.h 		mcl->args[1] = new_val.pte;
pte               469 arch/x86/include/asm/xen/hypercall.h 		mcl->args[2] = new_val.pte >> 32;
pte               315 arch/x86/include/asm/xen/page.h static inline unsigned long pte_mfn(pte_t pte)
pte               317 arch/x86/include/asm/xen/page.h 	return (pte.pte & XEN_PTE_MFN_MASK) >> PAGE_SHIFT;
pte               322 arch/x86/include/asm/xen/page.h 	pte_t pte;
pte               324 arch/x86/include/asm/xen/page.h 	pte.pte = ((phys_addr_t)page_nr << PAGE_SHIFT) |
pte               327 arch/x86/include/asm/xen/page.h 	return pte;
pte               330 arch/x86/include/asm/xen/page.h static inline pteval_t pte_val_ma(pte_t pte)
pte               332 arch/x86/include/asm/xen/page.h 	return pte.pte;
pte               337 arch/x86/include/asm/xen/page.h 	return (pte_t) { .pte = x };
pte               794 arch/x86/kernel/alternative.c 	pte_t pte, *ptep;
pte               838 arch/x86/kernel/alternative.c 	pte = mk_pte(pages[0], pgprot);
pte               839 arch/x86/kernel/alternative.c 	set_pte_at(poking_mm, poking_addr, ptep, pte);
pte               842 arch/x86/kernel/alternative.c 		pte = mk_pte(pages[1], pgprot);
pte               843 arch/x86/kernel/alternative.c 		set_pte_at(poking_mm, poking_addr + PAGE_SIZE, ptep + 1, pte);
pte               137 arch/x86/kernel/espfix_64.c 	pte_t pte, *pte_p;
pte               194 arch/x86/kernel/espfix_64.c 	pte = __pte(__pa(stack_page) | ((__PAGE_KERNEL_RO | _PAGE_ENC) & ptemask));
pte               196 arch/x86/kernel/espfix_64.c 		set_pte(&pte_p[n*PTE_STRIDE], pte);
pte                78 arch/x86/kernel/head32.c 	pte_t pte, *ptep;
pte                93 arch/x86/kernel/head32.c 	pte.pte = PTE_IDENT_ATTR;
pte                95 arch/x86/kernel/head32.c 	while ((pte.pte & PTE_PFN_MASK) < limit) {
pte               104 arch/x86/kernel/head32.c 			*ptep = pte;
pte               105 arch/x86/kernel/head32.c 			pte.pte += PAGE_SIZE;
pte               114 arch/x86/kernel/head32.c 	*ptr = (pte.pte & PTE_PFN_MASK) >> PAGE_SHIFT;
pte               232 arch/x86/kernel/ldt.c 		pte_t pte, *ptep;
pte               253 arch/x86/kernel/ldt.c 		pte = pfn_pte(pfn, pte_prot);
pte               254 arch/x86/kernel/ldt.c 		set_pte_at(mm, va, ptep, pte);
pte                92 arch/x86/kernel/machine_kexec_32.c 	pgd_t *pgd, pmd_t *pmd, pte_t *pte,
pte               107 arch/x86/kernel/machine_kexec_32.c 		set_pmd(pmd, __pmd(__pa(pte) | _PAGE_TABLE));
pte               108 arch/x86/kernel/machine_kexec_32.c 	pte = pte_offset_kernel(pmd, vaddr);
pte               109 arch/x86/kernel/machine_kexec_32.c 	set_pte(pte, pfn_pte(paddr >> PAGE_SHIFT, PAGE_KERNEL_EXEC));
pte               120 arch/x86/kernel/machine_kexec_64.c 	free_page((unsigned long)image->arch.pte);
pte               121 arch/x86/kernel/machine_kexec_64.c 	image->arch.pte = NULL;
pte               132 arch/x86/kernel/machine_kexec_64.c 	pte_t *pte;
pte               162 arch/x86/kernel/machine_kexec_64.c 		pte = (pte_t *)get_zeroed_page(GFP_KERNEL);
pte               163 arch/x86/kernel/machine_kexec_64.c 		if (!pte)
pte               165 arch/x86/kernel/machine_kexec_64.c 		image->arch.pte = pte;
pte               166 arch/x86/kernel/machine_kexec_64.c 		set_pmd(pmd, __pmd(__pa(pte) | _KERNPG_TABLE));
pte               168 arch/x86/kernel/machine_kexec_64.c 	pte = pte_offset_kernel(pmd, vaddr);
pte               173 arch/x86/kernel/machine_kexec_64.c 	set_pte(pte, pfn_pte(paddr >> PAGE_SHIFT, prot));
pte               111 arch/x86/kernel/tboot.c 	pte_t *pte;
pte               123 arch/x86/kernel/tboot.c 	pte = pte_alloc_map(&tboot_mm, pmd, vaddr);
pte               124 arch/x86/kernel/tboot.c 	if (!pte)
pte               126 arch/x86/kernel/tboot.c 	set_pte_at(&tboot_mm, vaddr, pte, pfn_pte(pfn, prot));
pte               127 arch/x86/kernel/tboot.c 	pte_unmap(pte);
pte               172 arch/x86/kernel/vm86_32.c 	pte_t *pte;
pte               193 arch/x86/kernel/vm86_32.c 	pte = pte_offset_map_lock(mm, pmd, 0xA0000, &ptl);
pte               195 arch/x86/kernel/vm86_32.c 		if (pte_present(*pte))
pte               196 arch/x86/kernel/vm86_32.c 			set_pte(pte, pte_wrprotect(*pte));
pte               197 arch/x86/kernel/vm86_32.c 		pte++;
pte               199 arch/x86/kernel/vm86_32.c 	pte_unmap_unlock(pte, ptl);
pte               609 arch/x86/kvm/mmu.c static int is_shadow_present_pte(u64 pte)
pte               611 arch/x86/kvm/mmu.c 	return (pte != 0) && !is_mmio_spte(pte);
pte               614 arch/x86/kvm/mmu.c static int is_large_pte(u64 pte)
pte               616 arch/x86/kvm/mmu.c 	return pte & PT_PAGE_SIZE_MASK;
pte               619 arch/x86/kvm/mmu.c static int is_last_spte(u64 pte, int level)
pte               623 arch/x86/kvm/mmu.c 	if (is_large_pte(pte))
pte               633 arch/x86/kvm/mmu.c static kvm_pfn_t spte_to_pfn(u64 pte)
pte               635 arch/x86/kvm/mmu.c 	return (pte & PT64_BASE_ADDR_MASK) >> PAGE_SHIFT;
pte              2053 arch/x86/kvm/mmu.c int kvm_set_spte_hva(struct kvm *kvm, unsigned long hva, pte_t pte)
pte              2055 arch/x86/kvm/mmu.c 	return kvm_handle_hva(kvm, hva, (unsigned long)&pte, kvm_set_pte_rmapp);
pte              2236 arch/x86/kvm/mmu.c 				 const void *pte)
pte              2751 arch/x86/kvm/mmu.c 	u64 pte;
pte              2754 arch/x86/kvm/mmu.c 	pte = *spte;
pte              2755 arch/x86/kvm/mmu.c 	if (is_shadow_present_pte(pte)) {
pte              2756 arch/x86/kvm/mmu.c 		if (is_last_spte(pte, sp->role.level)) {
pte              2758 arch/x86/kvm/mmu.c 			if (is_large_pte(pte))
pte              2761 arch/x86/kvm/mmu.c 			child = page_header(pte & PT64_BASE_ADDR_MASK);
pte              2767 arch/x86/kvm/mmu.c 	if (is_mmio_spte(pte))
pte              3174 arch/x86/kvm/mmu.c 			u64 pte = *sptep;
pte              3176 arch/x86/kvm/mmu.c 			child = page_header(pte & PT64_BASE_ADDR_MASK);
pte              4008 arch/x86/kvm/mmu.c __is_rsvd_bits_set(struct rsvd_bits_validate *rsvd_check, u64 pte, int level)
pte              4010 arch/x86/kvm/mmu.c 	int bit7 = (pte >> 7) & 1, low6 = pte & 0x3f;
pte              4012 arch/x86/kvm/mmu.c 	return (pte & rsvd_check->rsvd_bits_mask[bit7][level-1]) |
pte               138 arch/x86/kvm/mmu.h static inline int is_writable_pte(unsigned long pte)
pte               140 arch/x86/kvm/mmu.h 	return pte & PT_WRITABLE_MASK;
pte                83 arch/x86/kvm/mmutrace.h 	TP_PROTO(u64 pte, int level),
pte                84 arch/x86/kvm/mmutrace.h 	TP_ARGS(pte, level),
pte                87 arch/x86/kvm/mmutrace.h 		__field(__u64, pte)
pte                92 arch/x86/kvm/mmutrace.h 		__entry->pte = pte;
pte                96 arch/x86/kvm/mmutrace.h 	TP_printk("pte %llx level %u", __entry->pte, __entry->level)
pte                78 arch/x86/kvm/paging_tmpl.h #define gpte_to_gfn(pte) gpte_to_gfn_lvl((pte), PT_PAGE_TABLE_LEVEL)
pte               122 arch/x86/kvm/paging_tmpl.h static inline int FNAME(is_present_gpte)(unsigned long pte)
pte               125 arch/x86/kvm/paging_tmpl.h 	return pte & PT_PRESENT_MASK;
pte               127 arch/x86/kvm/paging_tmpl.h 	return pte & 7;
pte               226 arch/x86/kvm/paging_tmpl.h 	pt_element_t pte, orig_pte;
pte               236 arch/x86/kvm/paging_tmpl.h 		pte = orig_pte = walker->ptes[level - 1];
pte               240 arch/x86/kvm/paging_tmpl.h 		if (!(pte & PT_GUEST_ACCESSED_MASK)) {
pte               241 arch/x86/kvm/paging_tmpl.h 			trace_kvm_mmu_set_accessed_bit(table_gfn, index, sizeof(pte));
pte               242 arch/x86/kvm/paging_tmpl.h 			pte |= PT_GUEST_ACCESSED_MASK;
pte               245 arch/x86/kvm/paging_tmpl.h 				!(pte & PT_GUEST_DIRTY_MASK)) {
pte               246 arch/x86/kvm/paging_tmpl.h 			trace_kvm_mmu_set_dirty_bit(table_gfn, index, sizeof(pte));
pte               251 arch/x86/kvm/paging_tmpl.h 			pte |= PT_GUEST_DIRTY_MASK;
pte               253 arch/x86/kvm/paging_tmpl.h 		if (pte == orig_pte)
pte               272 arch/x86/kvm/paging_tmpl.h 		ret = FNAME(cmpxchg_gpte)(vcpu, mmu, ptep_user, index, orig_pte, pte);
pte               277 arch/x86/kvm/paging_tmpl.h 		walker->ptes[level - 1] = pte;
pte               286 arch/x86/kvm/paging_tmpl.h 	pte_t pte = {.pte = gpte};
pte               288 arch/x86/kvm/paging_tmpl.h 	pkeys = pte_flags_pkey(pte_flags(pte));
pte               301 arch/x86/kvm/paging_tmpl.h 	pt_element_t pte;
pte               321 arch/x86/kvm/paging_tmpl.h 	pte           = mmu->get_cr3(vcpu);
pte               327 arch/x86/kvm/paging_tmpl.h 		pte = mmu->get_pdptr(vcpu, (addr >> 30) & 3);
pte               328 arch/x86/kvm/paging_tmpl.h 		trace_kvm_mmu_paging_element(pte, walker->level);
pte               329 arch/x86/kvm/paging_tmpl.h 		if (!FNAME(is_present_gpte)(pte))
pte               355 arch/x86/kvm/paging_tmpl.h 		table_gfn = gpte_to_gfn(pte);
pte               388 arch/x86/kvm/paging_tmpl.h 		if (unlikely(__copy_from_user(&pte, ptep_user, sizeof(pte))))
pte               392 arch/x86/kvm/paging_tmpl.h 		trace_kvm_mmu_paging_element(pte, walker->level);
pte               398 arch/x86/kvm/paging_tmpl.h 		pte_access = pt_access & (pte ^ walk_nx_mask);
pte               400 arch/x86/kvm/paging_tmpl.h 		if (unlikely(!FNAME(is_present_gpte)(pte)))
pte               403 arch/x86/kvm/paging_tmpl.h 		if (unlikely(is_rsvd_bits_set(mmu, pte, walker->level))) {
pte               408 arch/x86/kvm/paging_tmpl.h 		walker->ptes[walker->level - 1] = pte;
pte               409 arch/x86/kvm/paging_tmpl.h 	} while (!is_last_gpte(mmu, walker->level, pte));
pte               411 arch/x86/kvm/paging_tmpl.h 	pte_pkey = FNAME(gpte_pkeys)(vcpu, pte);
pte               421 arch/x86/kvm/paging_tmpl.h 	gfn = gpte_to_gfn_lvl(pte, walker->level);
pte               425 arch/x86/kvm/paging_tmpl.h 		gfn += pse36_gfn_delta(pte);
pte               434 arch/x86/kvm/paging_tmpl.h 		FNAME(protect_clean_gpte)(mmu, &walker->pte_access, pte);
pte               441 arch/x86/kvm/paging_tmpl.h 		accessed_dirty &= pte >>
pte               453 arch/x86/kvm/paging_tmpl.h 		 __func__, (u64)pte, walker->pte_access, walker->pt_access);
pte               548 arch/x86/kvm/paging_tmpl.h 			      u64 *spte, const void *pte)
pte               550 arch/x86/kvm/paging_tmpl.h 	pt_element_t gpte = *(const pt_element_t *)pte;
pte                32 arch/x86/mm/cpu_entry_area.c 	pte_t pte = pfn_pte(pa >> PAGE_SHIFT, flags);
pte                43 arch/x86/mm/cpu_entry_area.c 		pte = pte_set_flags(pte, _PAGE_GLOBAL);
pte                45 arch/x86/mm/cpu_entry_area.c 	set_pte_vaddr(va, pte);
pte               361 arch/x86/mm/dump_pagetables.c 	pte_t *pte;
pte               366 arch/x86/mm/dump_pagetables.c 		pte = pte_offset_map(&addr, st->current_address);
pte               367 arch/x86/mm/dump_pagetables.c 		prot = pte_flags(*pte);
pte               370 arch/x86/mm/dump_pagetables.c 		pte_unmap(pte);
pte               298 arch/x86/mm/fault.c 	pte_t *pte;
pte               323 arch/x86/mm/fault.c 	pte = pte_offset_kernel(pmd, address);
pte               324 arch/x86/mm/fault.c 	pr_cont("*pte = %0*Lx ", sizeof(*pte) * 2, (u64)pte_val(*pte));
pte               359 arch/x86/mm/fault.c 	pte_t *pte;
pte               413 arch/x86/mm/fault.c 	pte = pte_offset_kernel(pmd, address);
pte               414 arch/x86/mm/fault.c 	if (!pte_present(*pte))
pte               453 arch/x86/mm/fault.c 	pte_t *pte;
pte               487 arch/x86/mm/fault.c 	pte = pte_offset_kernel(pmd, address);
pte               488 arch/x86/mm/fault.c 	if (bad_address(pte))
pte               491 arch/x86/mm/fault.c 	pr_cont("PTE %lx", pte_val(*pte));
pte               616 arch/x86/mm/fault.c 		pte_t *pte;
pte               621 arch/x86/mm/fault.c 		pte = lookup_address_in_pgd(pgd, address, &level);
pte               623 arch/x86/mm/fault.c 		if (pte && pte_present(*pte) && !pte_exec(*pte))
pte               626 arch/x86/mm/fault.c 		if (pte && pte_present(*pte) && pte_exec(*pte) &&
pte              1088 arch/x86/mm/fault.c static int spurious_kernel_fault_check(unsigned long error_code, pte_t *pte)
pte              1090 arch/x86/mm/fault.c 	if ((error_code & X86_PF_WRITE) && !pte_write(*pte))
pte              1093 arch/x86/mm/fault.c 	if ((error_code & X86_PF_INSTR) && !pte_exec(*pte))
pte              1127 arch/x86/mm/fault.c 	pte_t *pte;
pte              1168 arch/x86/mm/fault.c 	pte = pte_offset_kernel(pmd, address);
pte              1169 arch/x86/mm/fault.c 	if (!pte_present(*pte))
pte              1172 arch/x86/mm/fault.c 	ret = spurious_kernel_fault_check(error_code, pte);
pte                38 arch/x86/mm/hugetlbpage.c 	pte = huge_pte_offset(mm, address, vma_mmu_pagesize(vma));
pte                41 arch/x86/mm/hugetlbpage.c 	WARN_ON(!pte || pte_none(*pte));
pte                43 arch/x86/mm/hugetlbpage.c 	page = &pte_page(*pte)[vpfn % (HPAGE_SIZE/PAGE_SIZE)];
pte               157 arch/x86/mm/init_32.c static pte_t *__init page_table_kmap_check(pte_t *pte, pmd_t *pmd,
pte               180 arch/x86/mm/init_32.c 			set_pte(newpte + i, pte[i]);
pte               188 arch/x86/mm/init_32.c 		paravirt_release_pte(__pa(pte) >> PAGE_SHIFT);
pte               189 arch/x86/mm/init_32.c 		pte = newpte;
pte               193 arch/x86/mm/init_32.c 	       && lastpte && lastpte + PTRS_PER_PTE != pte);
pte               195 arch/x86/mm/init_32.c 	return pte;
pte               214 arch/x86/mm/init_32.c 	pte_t *pte = NULL;
pte               231 arch/x86/mm/init_32.c 			pte = page_table_kmap_check(one_page_table_init(pmd),
pte               232 arch/x86/mm/init_32.c 						    pmd, vaddr, pte, &adr);
pte               265 arch/x86/mm/init_32.c 	pte_t *pte;
pte               343 arch/x86/mm/init_32.c 			pte = one_page_table_init(pmd);
pte               346 arch/x86/mm/init_32.c 			pte += pte_ofs;
pte               348 arch/x86/mm/init_32.c 			     pte++, pfn++, pte_ofs++, addr += PAGE_SIZE) {
pte               361 arch/x86/mm/init_32.c 					set_pte(pte, pfn_pte(pfn, init_prot));
pte               364 arch/x86/mm/init_32.c 					set_pte(pte, pfn_pte(pfn, prot));
pte               421 arch/x86/mm/init_32.c 	pte_t *pte;
pte               430 arch/x86/mm/init_32.c 	pte = pte_offset_kernel(pmd, vaddr);
pte               431 arch/x86/mm/init_32.c 	pkmap_page_table = pte;
pte               478 arch/x86/mm/init_32.c 	pte_t *pte;
pte               508 arch/x86/mm/init_32.c 		pte = pte_offset_kernel(pmd, va);
pte               509 arch/x86/mm/init_32.c 		if (!pte_present(*pte))
pte               513 arch/x86/mm/init_32.c 				pfn, pmd, __pa(pmd), pte, __pa(pte));
pte               514 arch/x86/mm/init_32.c 		pte_clear(NULL, va, pte);
pte                75 arch/x86/mm/init_64.c DEFINE_POPULATE(pmd_populate_kernel, pmd, pte, init)
pte                90 arch/x86/mm/init_64.c DEFINE_ENTRY(pte, pte, init)
pte               282 arch/x86/mm/init_64.c 		pte_t *pte = (pte_t *) spp_getpage();
pte               283 arch/x86/mm/init_64.c 		pmd_populate_kernel(&init_mm, pmd, pte);
pte               284 arch/x86/mm/init_64.c 		if (pte != pte_offset_kernel(pmd, 0))
pte               293 arch/x86/mm/init_64.c 	pte_t *pte = fill_pte(pmd, vaddr);
pte               295 arch/x86/mm/init_64.c 	set_pte(pte, new_pte);
pte               453 arch/x86/mm/init_64.c 	pte_t *pte;
pte               456 arch/x86/mm/init_64.c 	pte = pte_page + pte_index(paddr);
pte               459 arch/x86/mm/init_64.c 	for (; i < PTRS_PER_PTE; i++, paddr = paddr_next, pte++) {
pte               467 arch/x86/mm/init_64.c 				set_pte_init(pte, __pte(0), init);
pte               477 arch/x86/mm/init_64.c 		if (!pte_none(*pte)) {
pte               484 arch/x86/mm/init_64.c 			pr_info("   pte=%p addr=%lx pte=%016lx\n", pte, paddr,
pte               485 arch/x86/mm/init_64.c 				pfn_pte(paddr >> PAGE_SHIFT, PAGE_KERNEL).pte);
pte               487 arch/x86/mm/init_64.c 		set_pte_init(pte, pfn_pte(paddr >> PAGE_SHIFT, prot), init);
pte               512 arch/x86/mm/init_64.c 		pte_t *pte;
pte               529 arch/x86/mm/init_64.c 				pte = (pte_t *)pmd_page_vaddr(*pmd);
pte               530 arch/x86/mm/init_64.c 				paddr_last = phys_pte_init(pte, paddr,
pte               569 arch/x86/mm/init_64.c 		pte = alloc_low_page();
pte               570 arch/x86/mm/init_64.c 		paddr_last = phys_pte_init(pte, paddr, paddr_end, new_prot, init);
pte               573 arch/x86/mm/init_64.c 		pmd_populate_kernel_init(&init_mm, pmd, pte, init);
pte               905 arch/x86/mm/init_64.c 	pte_t *pte;
pte               909 arch/x86/mm/init_64.c 		pte = pte_start + i;
pte               910 arch/x86/mm/init_64.c 		if (!pte_none(*pte))
pte               962 arch/x86/mm/init_64.c 	pte_t *pte;
pte               966 arch/x86/mm/init_64.c 	pte = pte_start + pte_index(addr);
pte               967 arch/x86/mm/init_64.c 	for (; addr < end; addr = next, pte++) {
pte               972 arch/x86/mm/init_64.c 		if (!pte_present(*pte))
pte               980 arch/x86/mm/init_64.c 		phys_addr = pte_val(*pte) + (addr & PAGE_MASK);
pte               990 arch/x86/mm/init_64.c 				free_pagetable(pte_page(*pte), 0);
pte               993 arch/x86/mm/init_64.c 			pte_clear(&init_mm, addr, pte);
pte              1011 arch/x86/mm/init_64.c 			page_addr = page_address(pte_page(*pte));
pte              1013 arch/x86/mm/init_64.c 				free_pagetable(pte_page(*pte), 0);
pte              1016 arch/x86/mm/init_64.c 				pte_clear(&init_mm, addr, pte);
pte              1348 arch/x86/mm/init_64.c 	pte_t *pte;
pte              1375 arch/x86/mm/init_64.c 	pte = pte_offset_kernel(pmd, addr);
pte              1376 arch/x86/mm/init_64.c 	if (pte_none(*pte))
pte              1379 arch/x86/mm/init_64.c 	return pfn_valid(pte_pfn(*pte));
pte              1549 arch/x86/mm/init_64.c 		pte_t *pte = NULL;
pte              1580 arch/x86/mm/init_64.c 			pte = pte_offset_kernel(pmd, addr);
pte              1581 arch/x86/mm/init_64.c 			if (pte_none(*pte))
pte              1583 arch/x86/mm/init_64.c 			get_page_bootmem(section_nr, pte_page(*pte),
pte               866 arch/x86/mm/ioremap.c 	pte_t *pte;
pte               872 arch/x86/mm/ioremap.c 	pte = early_ioremap_pte(addr);
pte               878 arch/x86/mm/ioremap.c 		set_pte(pte, pfn_pte(phys >> PAGE_SHIFT, flags));
pte               880 arch/x86/mm/ioremap.c 		pte_clear(&init_mm, addr, pte);
pte                42 arch/x86/mm/kasan_init_64.c 	pte_t *pte;
pte                61 arch/x86/mm/kasan_init_64.c 	pte = pte_offset_kernel(pmd, addr);
pte                66 arch/x86/mm/kasan_init_64.c 		if (!pte_none(*pte))
pte                71 arch/x86/mm/kasan_init_64.c 		set_pte_at(&init_mm, addr, pte, entry);
pte                72 arch/x86/mm/kasan_init_64.c 	} while (pte++, addr += PAGE_SIZE, addr != end);
pte               382 arch/x86/mm/kasan_init_64.c 		pte_t pte;
pte               388 arch/x86/mm/kasan_init_64.c 		pte = __pte(__pa(kasan_early_shadow_page) | pgprot_val(prot));
pte               389 arch/x86/mm/kasan_init_64.c 		set_pte(&kasan_early_shadow_pte[i], pte);
pte                77 arch/x86/mm/kmmio.c 	pte_t *pte = lookup_address(addr, &l);
pte                79 arch/x86/mm/kmmio.c 	if (!pte)
pte               114 arch/x86/mm/kmmio.c 	pte_t *pte = lookup_address(addr, &l);
pte               116 arch/x86/mm/kmmio.c 	if (!pte)
pte               141 arch/x86/mm/kmmio.c static void clear_pte_presence(pte_t *pte, bool clear, pteval_t *old)
pte               143 arch/x86/mm/kmmio.c 	pteval_t v = pte_val(*pte);
pte               147 arch/x86/mm/kmmio.c 		pte_clear(&init_mm, 0, pte);
pte               150 arch/x86/mm/kmmio.c 		set_pte_atomic(pte, __pte(*old));
pte               157 arch/x86/mm/kmmio.c 	pte_t *pte = lookup_address(f->addr, &level);
pte               159 arch/x86/mm/kmmio.c 	if (!pte) {
pte               166 arch/x86/mm/kmmio.c 		clear_pmd_presence((pmd_t *)pte, clear, &f->old_presence);
pte               169 arch/x86/mm/kmmio.c 		clear_pte_presence(pte, clear, &f->old_presence);
pte               237 arch/x86/mm/kmmio.c 	pte_t *pte = lookup_address(addr, &l);
pte               238 arch/x86/mm/kmmio.c 	if (!pte)
pte               446 arch/x86/mm/kmmio.c 	pte_t *pte;
pte               454 arch/x86/mm/kmmio.c 	pte = lookup_address(addr, &l);
pte               455 arch/x86/mm/kmmio.c 	if (!pte) {
pte               542 arch/x86/mm/kmmio.c 	pte_t *pte;
pte               544 arch/x86/mm/kmmio.c 	pte = lookup_address(addr, &l);
pte               545 arch/x86/mm/kmmio.c 	if (!pte)
pte               162 arch/x86/mm/mem_encrypt_identity.c 	pte_t *pte;
pte               170 arch/x86/mm/mem_encrypt_identity.c 		pte = ppd->pgtable_area;
pte               171 arch/x86/mm/mem_encrypt_identity.c 		memset(pte, 0, sizeof(*pte) * PTRS_PER_PTE);
pte               172 arch/x86/mm/mem_encrypt_identity.c 		ppd->pgtable_area += sizeof(*pte) * PTRS_PER_PTE;
pte               173 arch/x86/mm/mem_encrypt_identity.c 		set_pmd(pmd, __pmd(PMD_FLAGS | __pa(pte)));
pte               179 arch/x86/mm/mem_encrypt_identity.c 	pte = pte_offset_map(pmd, ppd->vaddr);
pte               180 arch/x86/mm/mem_encrypt_identity.c 	if (pte_none(*pte))
pte               181 arch/x86/mm/mem_encrypt_identity.c 		set_pte(pte, __pte(ppd->paddr | ppd->pte_flags));
pte                84 arch/x86/mm/mmio-mod.c 	pte_t *pte = lookup_address(address, &level);
pte                86 arch/x86/mm/mmio-mod.c 	if (!pte) {
pte                99 arch/x86/mm/mmio-mod.c 		(unsigned long long)pte_val(*pte),
pte               100 arch/x86/mm/mmio-mod.c 		(unsigned long long)pte_val(*pte) & _PAGE_PRESENT);
pte                40 arch/x86/mm/pageattr-test.c static int pte_testbit(pte_t pte)
pte                42 arch/x86/mm/pageattr-test.c 	return pte_flags(pte) & _PAGE_SOFTW1;
pte                61 arch/x86/mm/pageattr-test.c 		pte_t *pte;
pte                63 arch/x86/mm/pageattr-test.c 		pte = lookup_address(addr, &level);
pte                64 arch/x86/mm/pageattr-test.c 		if (!pte) {
pte                74 arch/x86/mm/pageattr-test.c 			if ((pte_val(*pte) & _PAGE_PRESENT) && !(pte_val(*pte) & _PAGE_PSE)) {
pte                77 arch/x86/mm/pageattr-test.c 					addr, level, (u64)pte_val(*pte));
pte                86 arch/x86/mm/pageattr-test.c 		if (!(pte_val(*pte) & _PAGE_NX)) {
pte               122 arch/x86/mm/pageattr-test.c 	pte_t *pte, pte0;
pte               149 arch/x86/mm/pageattr-test.c 		pte = NULL;
pte               153 arch/x86/mm/pageattr-test.c 			pte = lookup_address(addr[i] + k*PAGE_SIZE, &level);
pte               154 arch/x86/mm/pageattr-test.c 			if (!pte || pgprot_val(pte_pgprot(*pte)) == 0 ||
pte               155 arch/x86/mm/pageattr-test.c 			    !(pte_val(*pte) & _PAGE_PRESENT)) {
pte               160 arch/x86/mm/pageattr-test.c 				pte0 = *pte;
pte               162 arch/x86/mm/pageattr-test.c 				if (pgprot_val(pte_pgprot(*pte)) !=
pte               176 arch/x86/mm/pageattr-test.c 		if (!addr[i] || !pte || !k) {
pte               201 arch/x86/mm/pageattr-test.c 		pte = lookup_address(addr[i], &level);
pte               202 arch/x86/mm/pageattr-test.c 		if (!pte || !pte_testbit(*pte) || pte_huge(*pte)) {
pte               204 arch/x86/mm/pageattr-test.c 				pte ? (u64)pte_val(*pte) : 0ULL);
pte               221 arch/x86/mm/pageattr-test.c 		pte = lookup_address(addr[i], &level);
pte               222 arch/x86/mm/pageattr-test.c 		if (!pte) {
pte               232 arch/x86/mm/pageattr-test.c 		pte = lookup_address(addr[i], &level);
pte               233 arch/x86/mm/pageattr-test.c 		if (!pte || pte_testbit(*pte)) {
pte               235 arch/x86/mm/pageattr-test.c 				addr[i], pte ? (u64)pte_val(*pte) : 0ULL);
pte               369 arch/x86/mm/pageattr.c 		pte_t *pte = lookup_address(addr, &level);
pte               374 arch/x86/mm/pageattr.c 		if (pte && (pte_val(*pte) & _PAGE_PRESENT))
pte               674 arch/x86/mm/pageattr.c 	pte_t *pte;
pte               676 arch/x86/mm/pageattr.c 	pte = lookup_address(virt_addr, &level);
pte               677 arch/x86/mm/pageattr.c 	BUG_ON(!pte);
pte               686 arch/x86/mm/pageattr.c 		phys_addr = (phys_addr_t)pud_pfn(*(pud_t *)pte) << PAGE_SHIFT;
pte               690 arch/x86/mm/pageattr.c 		phys_addr = (phys_addr_t)pmd_pfn(*(pmd_t *)pte) << PAGE_SHIFT;
pte               694 arch/x86/mm/pageattr.c 		phys_addr = (phys_addr_t)pte_pfn(*pte) << PAGE_SHIFT;
pte               705 arch/x86/mm/pageattr.c static void __set_pmd_pte(pte_t *kpte, unsigned long address, pte_t pte)
pte               708 arch/x86/mm/pageattr.c 	set_pte_atomic(kpte, pte);
pte               723 arch/x86/mm/pageattr.c 			set_pte_atomic((pte_t *)pmd, pte);
pte               904 arch/x86/mm/pageattr.c static void split_set_pte(struct cpa_data *cpa, pte_t *pte, unsigned long pfn,
pte               937 arch/x86/mm/pageattr.c 	set_pte(pte, pfn_pte(pfn, ref_prot));
pte              1064 arch/x86/mm/pageattr.c static bool try_to_free_pte_page(pte_t *pte)
pte              1069 arch/x86/mm/pageattr.c 		if (!pte_none(pte[i]))
pte              1072 arch/x86/mm/pageattr.c 	free_page((unsigned long)pte);
pte              1090 arch/x86/mm/pageattr.c 	pte_t *pte = pte_offset_kernel(pmd, start);
pte              1093 arch/x86/mm/pageattr.c 		set_pte(pte, __pte(0));
pte              1096 arch/x86/mm/pageattr.c 		pte++;
pte              1203 arch/x86/mm/pageattr.c 	pte_t *pte = (pte_t *)get_zeroed_page(GFP_KERNEL);
pte              1204 arch/x86/mm/pageattr.c 	if (!pte)
pte              1207 arch/x86/mm/pageattr.c 	set_pmd(pmd, __pmd(__pa(pte) | _KERNPG_TABLE));
pte              1225 arch/x86/mm/pageattr.c 	pte_t *pte;
pte              1227 arch/x86/mm/pageattr.c 	pte = pte_offset_kernel(pmd, start);
pte              1232 arch/x86/mm/pageattr.c 		set_pte(pte, pfn_pte(cpa->pfn, pgprot));
pte              1236 arch/x86/mm/pageattr.c 		pte++;
pte              2201 arch/x86/mm/pageattr.c 	pte_t *pte;
pte              2206 arch/x86/mm/pageattr.c 	pte = lookup_address((unsigned long)page_address(page), &level);
pte              2207 arch/x86/mm/pageattr.c 	return (pte_val(*pte) & _PAGE_PRESENT);
pte                46 arch/x86/mm/pgtable.c void ___pte_free_tlb(struct mmu_gather *tlb, struct page *pte)
pte                48 arch/x86/mm/pgtable.c 	pgtable_pte_page_dtor(pte);
pte                49 arch/x86/mm/pgtable.c 	paravirt_release_pte(page_to_pfn(pte));
pte                50 arch/x86/mm/pgtable.c 	paravirt_tlb_remove_table(tlb, pte);
pte               541 arch/x86/mm/pgtable.c 					 (unsigned long *) &ptep->pte);
pte               625 arch/x86/mm/pgtable.c void __native_set_fixmap(enum fixed_addresses idx, pte_t pte)
pte               642 arch/x86/mm/pgtable.c 	set_pte_vaddr(address, pte);
pte               802 arch/x86/mm/pgtable.c 	pte_t *pte;
pte               823 arch/x86/mm/pgtable.c 			pte = (pte_t *)pmd_page_vaddr(pmd_sv[i]);
pte               824 arch/x86/mm/pgtable.c 			free_page((unsigned long)pte);
pte               844 arch/x86/mm/pgtable.c 	pte_t *pte;
pte               846 arch/x86/mm/pgtable.c 	pte = (pte_t *)pmd_page_vaddr(*pmd);
pte               852 arch/x86/mm/pgtable.c 	free_page((unsigned long)pte);
pte                34 arch/x86/mm/pgtable_32.c 	pte_t *pte;
pte                56 arch/x86/mm/pgtable_32.c 	pte = pte_offset_kernel(pmd, vaddr);
pte                58 arch/x86/mm/pgtable_32.c 		set_pte_at(&init_mm, vaddr, pte, pteval);
pte                60 arch/x86/mm/pgtable_32.c 		pte_clear(&init_mm, vaddr, pte);
pte               249 arch/x86/mm/pti.c 	pte_t *pte;
pte               269 arch/x86/mm/pti.c 	pte = pte_offset_kernel(pmd, address);
pte               270 arch/x86/mm/pti.c 	if (pte_flags(*pte) & _PAGE_USER) {
pte               274 arch/x86/mm/pti.c 	return pte;
pte               280 arch/x86/mm/pti.c 	pte_t *pte, *target_pte;
pte               283 arch/x86/mm/pti.c 	pte = lookup_address(VSYSCALL_ADDR, &level);
pte               284 arch/x86/mm/pti.c 	if (!pte || WARN_ON(level != PG_LEVEL_4K) || pte_none(*pte))
pte               291 arch/x86/mm/pti.c 	*target_pte = *pte;
pte               314 arch/x86/mm/pti.c 		pte_t *pte, *target_pte;
pte               383 arch/x86/mm/pti.c 			pte = pte_offset_kernel(pmd, addr);
pte               384 arch/x86/mm/pti.c 			if (pte_none(*pte)) {
pte               390 arch/x86/mm/pti.c 			if (WARN_ON(!(pte_flags(*pte) & _PAGE_PRESENT)))
pte               400 arch/x86/mm/pti.c 				*pte = pte_set_flags(*pte, _PAGE_GLOBAL);
pte               403 arch/x86/mm/pti.c 			*target_pte = *pte;
pte               217 arch/x86/power/hibernate.c 	pte_t *pte;
pte               243 arch/x86/power/hibernate.c 	pte = pte_offset_kernel(pmd, relocated_restore_code);
pte               244 arch/x86/power/hibernate.c 	set_pte(pte, __pte(pte_val(*pte) & ~_PAGE_NX));
pte                86 arch/x86/power/hibernate_32.c 	pte_t *pte;
pte               115 arch/x86/power/hibernate_32.c 				pte = resume_one_page_table_init(pmd);
pte               116 arch/x86/power/hibernate_32.c 				if (!pte)
pte               119 arch/x86/power/hibernate_32.c 				max_pte = pte + PTRS_PER_PTE;
pte               120 arch/x86/power/hibernate_32.c 				for (; pte < max_pte; pte++, pfn++) {
pte               124 arch/x86/power/hibernate_32.c 					set_pte(pte, pfn_pte(pfn, PAGE_KERNEL_EXEC));
pte               149 arch/x86/power/hibernate_32.c 	pte_t *pte;
pte               161 arch/x86/power/hibernate_32.c 		pte = resume_one_page_table_init(pmd);
pte               162 arch/x86/power/hibernate_32.c 		if (!pte)
pte               164 arch/x86/power/hibernate_32.c 		set_pte(pte + pte_index(restore_jump_address),
pte               352 arch/x86/xen/enlighten_pv.c 	pte_t pte;
pte               363 arch/x86/xen/enlighten_pv.c 	pte = pfn_pte(pfn, prot);
pte               389 arch/x86/xen/enlighten_pv.c 	if (HYPERVISOR_update_va_mapping((unsigned long)v, pte, 0))
pte               396 arch/x86/xen/enlighten_pv.c 			if (HYPERVISOR_update_va_mapping((unsigned long)av, pte, 0))
pte               492 arch/x86/xen/enlighten_pv.c 	pte_t pte;
pte               501 arch/x86/xen/enlighten_pv.c 	pte = pfn_pte(pfn, PAGE_KERNEL_RO);
pte               503 arch/x86/xen/enlighten_pv.c 	if (HYPERVISOR_update_va_mapping((unsigned long)va, pte, 0))
pte                22 arch/x86/xen/mmu.c 	pte_t *pte;
pte                34 arch/x86/xen/mmu.c 	pte = lookup_address(address, &level);
pte                35 arch/x86/xen/mmu.c 	BUG_ON(pte == NULL);
pte                37 arch/x86/xen/mmu.c 	return XMADDR(((phys_addr_t)pte_mfn(*pte) << PAGE_SHIFT) + offset);
pte                22 arch/x86/xen/mmu.h 				  pte_t *ptep, pte_t pte);
pte               138 arch/x86/xen/mmu_pv.c 	pte_t *pte, ptev;
pte               142 arch/x86/xen/mmu_pv.c 	pte = lookup_address(address, &level);
pte               143 arch/x86/xen/mmu_pv.c 	if (pte == NULL)
pte               146 arch/x86/xen/mmu_pv.c 	ptev = pte_wrprotect(*pte);
pte               154 arch/x86/xen/mmu_pv.c 	pte_t *pte, ptev;
pte               158 arch/x86/xen/mmu_pv.c 	pte = lookup_address(address, &level);
pte               159 arch/x86/xen/mmu_pv.c 	if (pte == NULL)
pte               162 arch/x86/xen/mmu_pv.c 	ptev = pte_mkwrite(*pte);
pte               318 arch/x86/xen/mmu_pv.c 				 pte_t *ptep, pte_t pte)
pte               322 arch/x86/xen/mmu_pv.c 	trace_xen_mmu_ptep_modify_prot_commit(vma->vm_mm, addr, ptep, pte);
pte               326 arch/x86/xen/mmu_pv.c 	u.val = pte_val_ma(pte);
pte               375 arch/x86/xen/mmu_pv.c __visible pteval_t xen_pte_val(pte_t pte)
pte               377 arch/x86/xen/mmu_pv.c 	pteval_t pteval = pte.pte;
pte               389 arch/x86/xen/mmu_pv.c __visible pte_t xen_make_pte(pteval_t pte)
pte               391 arch/x86/xen/mmu_pv.c 	pte = pte_pfn_to_mfn(pte);
pte               393 arch/x86/xen/mmu_pv.c 	return native_make_pte(pte);
pte               443 arch/x86/xen/mmu_pv.c static void xen_set_pte_atomic(pte_t *ptep, pte_t pte)
pte               445 arch/x86/xen/mmu_pv.c 	trace_xen_mmu_set_pte_atomic(ptep, pte);
pte               446 arch/x86/xen/mmu_pv.c 	__xen_set_pte(ptep, pte);
pte              1540 arch/x86/xen/mmu_pv.c __visible pte_t xen_make_pte_init(pteval_t pte)
pte              1551 arch/x86/xen/mmu_pv.c 	pfn = (pte & PTE_PFN_MASK) >> PAGE_SHIFT;
pte              1555 arch/x86/xen/mmu_pv.c 		pte &= ~_PAGE_RW;
pte              1557 arch/x86/xen/mmu_pv.c 	pte = pte_pfn_to_mfn(pte);
pte              1558 arch/x86/xen/mmu_pv.c 	return native_make_pte(pte);
pte              1562 arch/x86/xen/mmu_pv.c static void __init xen_set_pte_init(pte_t *ptep, pte_t pte)
pte              1566 arch/x86/xen/mmu_pv.c 	if (pte_mfn(pte) != INVALID_P2M_ENTRY
pte              1568 arch/x86/xen/mmu_pv.c 		pte = __pte_ma(((pte_val_ma(*ptep) & _PAGE_RW) | ~_PAGE_RW) &
pte              1569 arch/x86/xen/mmu_pv.c 			       pte_val_ma(pte));
pte              1571 arch/x86/xen/mmu_pv.c 	__xen_set_pte(ptep, pte);
pte              1765 arch/x86/xen/mmu_pv.c 	pte_t pte = pfn_pte(pfn, prot);
pte              1767 arch/x86/xen/mmu_pv.c 	if (HYPERVISOR_update_va_mapping((unsigned long)addr, pte, flags))
pte              1805 arch/x86/xen/mmu_pv.c 			pte_t pte;
pte              1813 arch/x86/xen/mmu_pv.c 			pte = pfn_pte(pfn, PAGE_KERNEL_EXEC);
pte              1814 arch/x86/xen/mmu_pv.c 			pte_page[pteidx] = pte;
pte              1843 arch/x86/xen/mmu_pv.c 	pte_t *pte = v;
pte              1849 arch/x86/xen/mmu_pv.c 		pte[i] = xen_make_pte(pte[i].pte);
pte              2018 arch/x86/xen/mmu_pv.c 	pte_t pte;
pte              2043 arch/x86/xen/mmu_pv.c 	pte = native_make_pte(xen_read_phys_ulong(pa + pte_index(vaddr) *
pte              2044 arch/x86/xen/mmu_pv.c 						       sizeof(pte)));
pte              2045 arch/x86/xen/mmu_pv.c 	if (!pte_present(pte))
pte              2047 arch/x86/xen/mmu_pv.c 	pa = pte_pfn(pte) << PAGE_SHIFT;
pte              2297 arch/x86/xen/mmu_pv.c 	pte_t pte;
pte              2312 arch/x86/xen/mmu_pv.c 		pte = pfn_pte(phys, prot);
pte              2317 arch/x86/xen/mmu_pv.c 		pte = pfn_pte(PFN_DOWN(__pa(dummy_mapping)), PAGE_KERNEL);
pte              2327 arch/x86/xen/mmu_pv.c 		pte = pfn_pte(PFN_DOWN(__pa(dummy_mapping)), PAGE_KERNEL);
pte              2334 arch/x86/xen/mmu_pv.c 		pte = mfn_pte(phys, prot);
pte              2339 arch/x86/xen/mmu_pv.c 		pte = mfn_pte(phys, prot);
pte              2343 arch/x86/xen/mmu_pv.c 	__native_set_fixmap(idx, pte);
pte              2350 arch/x86/xen/mmu_pv.c 		set_pte_vaddr_pud(level3_user_vsyscall, vaddr, pte);
pte              2694 arch/x86/xen/mmu_pv.c 	pte_t pte = pte_mkspecial(mfn_pte(*rmd->pfn, rmd->prot));
pte              2709 arch/x86/xen/mmu_pv.c 	rmd->mmu_update->val = pte_val_ma(pte);
pte               703 arch/x86/xen/p2m.c 	pte_t *pte;
pte               723 arch/x86/xen/p2m.c 			pte = (pte_t *)(mfn_to_virt(PFN_DOWN(map_ops[i].host_addr)) +
pte               725 arch/x86/xen/p2m.c 			mfn = pte_mfn(*pte);
pte                97 arch/xtensa/include/asm/page.h typedef struct { unsigned long pte; } pte_t;		/* page table entry */
pte               102 arch/xtensa/include/asm/page.h #define pte_val(x)	((x).pte)
pte                51 arch/xtensa/include/asm/pgalloc.h 	pte_t *pte;
pte                54 arch/xtensa/include/asm/pgalloc.h 	pte = pte_alloc_one_kernel(mm);
pte                55 arch/xtensa/include/asm/pgalloc.h 	if (!pte)
pte                57 arch/xtensa/include/asm/pgalloc.h 	page = virt_to_page(pte);
pte                65 arch/xtensa/include/asm/pgalloc.h static inline void pte_free_kernel(struct mm_struct *mm, pte_t *pte)
pte                67 arch/xtensa/include/asm/pgalloc.h 	free_page((unsigned long)pte);
pte                70 arch/xtensa/include/asm/pgalloc.h static inline void pte_free(struct mm_struct *mm, pgtable_t pte)
pte                72 arch/xtensa/include/asm/pgalloc.h 	pgtable_pte_page_dtor(pte);
pte                73 arch/xtensa/include/asm/pgalloc.h 	__free_page(pte);
pte               251 arch/xtensa/include/asm/pgtable.h # define pte_none(pte)	 (pte_val(pte) == (_PAGE_CA_INVALID | _PAGE_USER))
pte               253 arch/xtensa/include/asm/pgtable.h # define pte_present(pte) ((pte_val(pte) & _PAGE_CA_MASK) != _PAGE_CA_INVALID)
pte               255 arch/xtensa/include/asm/pgtable.h # define pte_present(pte)						\
pte               256 arch/xtensa/include/asm/pgtable.h 	(((pte_val(pte) & _PAGE_CA_MASK) != _PAGE_CA_INVALID)		\
pte               257 arch/xtensa/include/asm/pgtable.h 	 || ((pte_val(pte) & _PAGE_ATTRIB_MASK) == _PAGE_NONE))
pte               267 arch/xtensa/include/asm/pgtable.h static inline int pte_write(pte_t pte) { return pte_val(pte) & _PAGE_WRITABLE; }
pte               268 arch/xtensa/include/asm/pgtable.h static inline int pte_dirty(pte_t pte) { return pte_val(pte) & _PAGE_DIRTY; }
pte               269 arch/xtensa/include/asm/pgtable.h static inline int pte_young(pte_t pte) { return pte_val(pte) & _PAGE_ACCESSED; }
pte               270 arch/xtensa/include/asm/pgtable.h static inline int pte_special(pte_t pte) { return 0; }
pte               272 arch/xtensa/include/asm/pgtable.h static inline pte_t pte_wrprotect(pte_t pte)	
pte               273 arch/xtensa/include/asm/pgtable.h 	{ pte_val(pte) &= ~(_PAGE_WRITABLE | _PAGE_HW_WRITE); return pte; }
pte               274 arch/xtensa/include/asm/pgtable.h static inline pte_t pte_mkclean(pte_t pte)
pte               275 arch/xtensa/include/asm/pgtable.h 	{ pte_val(pte) &= ~(_PAGE_DIRTY | _PAGE_HW_WRITE); return pte; }
pte               276 arch/xtensa/include/asm/pgtable.h static inline pte_t pte_mkold(pte_t pte)
pte               277 arch/xtensa/include/asm/pgtable.h 	{ pte_val(pte) &= ~_PAGE_ACCESSED; return pte; }
pte               278 arch/xtensa/include/asm/pgtable.h static inline pte_t pte_mkdirty(pte_t pte)
pte               279 arch/xtensa/include/asm/pgtable.h 	{ pte_val(pte) |= _PAGE_DIRTY; return pte; }
pte               280 arch/xtensa/include/asm/pgtable.h static inline pte_t pte_mkyoung(pte_t pte)
pte               281 arch/xtensa/include/asm/pgtable.h 	{ pte_val(pte) |= _PAGE_ACCESSED; return pte; }
pte               282 arch/xtensa/include/asm/pgtable.h static inline pte_t pte_mkwrite(pte_t pte)
pte               283 arch/xtensa/include/asm/pgtable.h 	{ pte_val(pte) |= _PAGE_WRITABLE; return pte; }
pte               284 arch/xtensa/include/asm/pgtable.h static inline pte_t pte_mkspecial(pte_t pte)
pte               285 arch/xtensa/include/asm/pgtable.h 	{ return pte; }
pte               294 arch/xtensa/include/asm/pgtable.h #define pte_pfn(pte)		(pte_val(pte) >> PAGE_SHIFT)
pte               300 arch/xtensa/include/asm/pgtable.h static inline pte_t pte_modify(pte_t pte, pgprot_t newprot)
pte               302 arch/xtensa/include/asm/pgtable.h 	return __pte((pte_val(pte) & _PAGE_CHG_MASK) | pgprot_val(newprot));
pte               344 arch/xtensa/include/asm/pgtable.h 	pte_t pte = *ptep;
pte               345 arch/xtensa/include/asm/pgtable.h 	if (!pte_young(pte))
pte               347 arch/xtensa/include/asm/pgtable.h 	update_pte(ptep, pte_mkold(pte));
pte               354 arch/xtensa/include/asm/pgtable.h 	pte_t pte = *ptep;
pte               356 arch/xtensa/include/asm/pgtable.h 	return pte;
pte               362 arch/xtensa/include/asm/pgtable.h 	pte_t pte = *ptep;
pte               363 arch/xtensa/include/asm/pgtable.h 	update_pte(ptep, pte_wrprotect(pte));
pte               382 arch/xtensa/include/asm/pgtable.h #define pte_unmap(pte)		do { } while (0)
pte               396 arch/xtensa/include/asm/pgtable.h #define __pte_to_swp_entry(pte)	((swp_entry_t) { pte_val(pte) })
pte                19 arch/xtensa/include/asm/tlb.h #define __pte_free_tlb(tlb, pte, address)	pte_free((tlb)->mm, pte)
pte                46 arch/xtensa/mm/kasan_init.c 	pte_t *pte = memblock_alloc(n_pages * sizeof(pte_t), PAGE_SIZE);
pte                48 arch/xtensa/mm/kasan_init.c 	if (!pte)
pte                66 arch/xtensa/mm/kasan_init.c 			set_pte(pte + j, pfn_pte(PHYS_PFN(phys), PAGE_KERNEL));
pte                70 arch/xtensa/mm/kasan_init.c 	for (i = 0; i < n_pmds ; ++i, pte += PTRS_PER_PTE)
pte                71 arch/xtensa/mm/kasan_init.c 		set_pmd(pmd + i, __pmd((unsigned long)pte));
pte                26 arch/xtensa/mm/mmu.c 	pte_t *pte;
pte                34 arch/xtensa/mm/mmu.c 	pte = memblock_alloc_low(n_pages * sizeof(pte_t), PAGE_SIZE);
pte                35 arch/xtensa/mm/mmu.c 	if (!pte)
pte                40 arch/xtensa/mm/mmu.c 		pte_clear(NULL, 0, pte + i);
pte                43 arch/xtensa/mm/mmu.c 		pte_t *cur_pte = pte + i;
pte                51 arch/xtensa/mm/mmu.c 	return pte;
pte               173 arch/xtensa/mm/tlb.c 	pte_t *pte;
pte               183 arch/xtensa/mm/tlb.c 	pte = pte_offset_map(pmd, vaddr);
pte               184 arch/xtensa/mm/tlb.c 	if (!pte)
pte               186 arch/xtensa/mm/tlb.c 	return pte_val(*pte);
pte               222 arch/xtensa/mm/tlb.c 	unsigned pte = get_pte_for_vaddr(vpn);
pte               236 arch/xtensa/mm/tlb.c 		if ((pte ^ r1) & PAGE_MASK) {
pte               238 arch/xtensa/mm/tlb.c 					dtlb ? 'D' : 'I', w, e, r0, r1, pte);
pte               239 arch/xtensa/mm/tlb.c 			if (pte == 0 || !pte_present(__pte(pte))) {
pte               277 block/partitions/efi.c 	gpt_entry *pte;
pte               286 block/partitions/efi.c 	pte = kmalloc(count, GFP_KERNEL);
pte               287 block/partitions/efi.c 	if (!pte)
pte               291 block/partitions/efi.c 			(u8 *) pte, count) < count) {
pte               292 block/partitions/efi.c 		kfree(pte);
pte               293 block/partitions/efi.c                 pte=NULL;
pte               296 block/partitions/efi.c 	return pte;
pte               463 block/partitions/efi.c is_pte_valid(const gpt_entry *pte, const u64 lastlba)
pte               465 block/partitions/efi.c 	if ((!efi_guidcmp(pte->partition_type_guid, NULL_GUID)) ||
pte               466 block/partitions/efi.c 	    le64_to_cpu(pte->starting_lba) > lastlba         ||
pte               467 block/partitions/efi.c 	    le64_to_cpu(pte->ending_lba)   > lastlba)
pte                51 drivers/char/agp/amd64-agp.c 	u32 pte;
pte                87 drivers/char/agp/amd64-agp.c 		pte = (tmp & 0x000000ff00000000ULL) >> 28;
pte                88 drivers/char/agp/amd64-agp.c 		pte |=(tmp & 0x00000000fffff000ULL);
pte                89 drivers/char/agp/amd64-agp.c 		pte |= GPTE_VALID | GPTE_COHERENT;
pte                91 drivers/char/agp/amd64-agp.c 		writel(pte, agp_bridge->gatt_table+j);
pte               201 drivers/gpu/drm/amd/amdgpu/amdgpu_vm_sdma.c 	uint64_t *pte;
pte               248 drivers/gpu/drm/amd/amdgpu/amdgpu_vm_sdma.c 		pte = (uint64_t *)&(p->job->ibs->ptr[p->num_dw_left]);
pte               250 drivers/gpu/drm/amd/amdgpu/amdgpu_vm_sdma.c 			pte[i] = amdgpu_vm_map_gart(p->pages_addr, addr);
pte               251 drivers/gpu/drm/amd/amdgpu/amdgpu_vm_sdma.c 			pte[i] |= flags;
pte               142 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.c 	const struct pte_setting *pte = &pte_settings[mi_tiling][mi_bpp];
pte               144 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.c 	unsigned int page_width = log_2(pte->page_width);
pte               145 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.c 	unsigned int page_height = log_2(pte->page_height);
pte               147 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.c 			pte->min_pte_before_flip_vert_scan :
pte               148 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.c 			pte->min_pte_before_flip_horiz_scan;
pte               159 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.c 			DVMM_PTE_REQ_PER_CHUNK, pte->pte_req_per_chunk,
pte               570 drivers/gpu/drm/amd/display/dc/dce110/dce110_mem_input_v.c 	const unsigned int *pte = get_dvmm_hw_setting(tiling_info, format, false);
pte               577 drivers/gpu/drm/amd/display/dc/dce110/dce110_mem_input_v.c 	unsigned int temp_page_width = pte[1];
pte               578 drivers/gpu/drm/amd/display/dc/dce110/dce110_mem_input_v.c 	unsigned int temp_page_height = pte[2];
pte               598 drivers/gpu/drm/amd/display/dc/dce110/dce110_mem_input_v.c 		min_pte_before_flip = pte[4];
pte               602 drivers/gpu/drm/amd/display/dc/dce110/dce110_mem_input_v.c 		min_pte_before_flip = pte[3];
pte               620 drivers/gpu/drm/amd/display/dc/dce110/dce110_mem_input_v.c 	set_reg_field_value(value, pte[5], UNP_DVMM_PTE_ARB_CONTROL, DVMM_PTE_REQ_PER_CHUNK);
pte                79 drivers/gpu/drm/gma500/gtt.c 	u32 pte;
pte               100 drivers/gpu/drm/gma500/gtt.c 		pte = psb_gtt_mask_pte(page_to_pfn(r->pages[i]),
pte               102 drivers/gpu/drm/gma500/gtt.c 		iowrite32(pte, gtt_slot++);
pte               105 drivers/gpu/drm/gma500/gtt.c 		pte = psb_gtt_mask_pte(page_to_pfn(r->pages[i]),
pte               107 drivers/gpu/drm/gma500/gtt.c 		iowrite32(pte, gtt_slot++);
pte               128 drivers/gpu/drm/gma500/gtt.c 	u32 pte;
pte               134 drivers/gpu/drm/gma500/gtt.c 	pte = psb_gtt_mask_pte(page_to_pfn(dev_priv->scratch_page),
pte               138 drivers/gpu/drm/gma500/gtt.c 		iowrite32(pte, gtt_slot++);
pte               156 drivers/gpu/drm/gma500/gtt.c 	u32 pte;
pte               174 drivers/gpu/drm/gma500/gtt.c 		pte = psb_gtt_mask_pte(page_to_pfn(r->pages[i]),
pte               176 drivers/gpu/drm/gma500/gtt.c 		iowrite32(pte, gtt_slot++);
pte               179 drivers/gpu/drm/gma500/gtt.c 		pte = psb_gtt_mask_pte(page_to_pfn(r->pages[i]),
pte               181 drivers/gpu/drm/gma500/gtt.c 		iowrite32(pte, gtt_slot++);
pte               416 drivers/gpu/drm/gma500/gtt.c 	uint32_t pte;
pte               533 drivers/gpu/drm/gma500/gtt.c 		pte = psb_gtt_mask_pte(pfn_base + i, PSB_MMU_CACHED_MEMORY);
pte               534 drivers/gpu/drm/gma500/gtt.c 		iowrite32(pte, dev_priv->gtt_map + i);
pte               542 drivers/gpu/drm/gma500/gtt.c 	pte = psb_gtt_mask_pte(pfn_base, PSB_MMU_CACHED_MEMORY);
pte               544 drivers/gpu/drm/gma500/gtt.c 		iowrite32(pte, dev_priv->gtt_map + i);
pte               397 drivers/gpu/drm/gma500/mmu.c 				   uint32_t pte)
pte               399 drivers/gpu/drm/gma500/mmu.c 	pt->v[psb_mmu_pt_index(addr)] = pte;
pte               655 drivers/gpu/drm/gma500/mmu.c 	uint32_t pte;
pte               675 drivers/gpu/drm/gma500/mmu.c 			pte = psb_mmu_mask_pte(start_pfn++, type);
pte               676 drivers/gpu/drm/gma500/mmu.c 			psb_mmu_set_pte(pt, addr, pte);
pte               704 drivers/gpu/drm/gma500/mmu.c 	uint32_t pte;
pte               737 drivers/gpu/drm/gma500/mmu.c 				pte = psb_mmu_mask_pte(page_to_pfn(*pages++),
pte               739 drivers/gpu/drm/gma500/mmu.c 				psb_mmu_set_pte(pt, addr, pte);
pte                37 drivers/gpu/drm/i915/gvt/dmabuf.c #define GEN8_DECODE_PTE(pte) (pte & GENMASK_ULL(63, 12))
pte               293 drivers/gpu/drm/i915/gvt/gtt.c 		unsigned long index, u64 pte)
pte               297 drivers/gpu/drm/i915/gvt/gtt.c 	writeq(pte, addr);
pte               493 drivers/gpu/drm/i915/gvt/gtt.c DEFINE_PPGTT_GMA_TO_INDEX(gen8, pte, (gma >> 12 & 0x1ff));
pte               202 drivers/gpu/drm/i915/i915_gem_gtt.c 	gen8_pte_t pte = addr | _PAGE_PRESENT | _PAGE_RW;
pte               205 drivers/gpu/drm/i915/i915_gem_gtt.c 		pte &= ~_PAGE_RW;
pte               209 drivers/gpu/drm/i915/i915_gem_gtt.c 		pte |= PPAT_UNCACHED;
pte               212 drivers/gpu/drm/i915/i915_gem_gtt.c 		pte |= PPAT_DISPLAY_ELLC;
pte               215 drivers/gpu/drm/i915/i915_gem_gtt.c 		pte |= PPAT_CACHED;
pte               219 drivers/gpu/drm/i915/i915_gem_gtt.c 	return pte;
pte               238 drivers/gpu/drm/i915/i915_gem_gtt.c 	gen6_pte_t pte = GEN6_PTE_VALID;
pte               239 drivers/gpu/drm/i915/i915_gem_gtt.c 	pte |= GEN6_PTE_ADDR_ENCODE(addr);
pte               244 drivers/gpu/drm/i915/i915_gem_gtt.c 		pte |= GEN6_PTE_CACHE_LLC;
pte               247 drivers/gpu/drm/i915/i915_gem_gtt.c 		pte |= GEN6_PTE_UNCACHED;
pte               253 drivers/gpu/drm/i915/i915_gem_gtt.c 	return pte;
pte               260 drivers/gpu/drm/i915/i915_gem_gtt.c 	gen6_pte_t pte = GEN6_PTE_VALID;
pte               261 drivers/gpu/drm/i915/i915_gem_gtt.c 	pte |= GEN6_PTE_ADDR_ENCODE(addr);
pte               265 drivers/gpu/drm/i915/i915_gem_gtt.c 		pte |= GEN7_PTE_CACHE_L3_LLC;
pte               268 drivers/gpu/drm/i915/i915_gem_gtt.c 		pte |= GEN6_PTE_CACHE_LLC;
pte               271 drivers/gpu/drm/i915/i915_gem_gtt.c 		pte |= GEN6_PTE_UNCACHED;
pte               277 drivers/gpu/drm/i915/i915_gem_gtt.c 	return pte;
pte               284 drivers/gpu/drm/i915/i915_gem_gtt.c 	gen6_pte_t pte = GEN6_PTE_VALID;
pte               285 drivers/gpu/drm/i915/i915_gem_gtt.c 	pte |= GEN6_PTE_ADDR_ENCODE(addr);
pte               288 drivers/gpu/drm/i915/i915_gem_gtt.c 		pte |= BYT_PTE_WRITEABLE;
pte               291 drivers/gpu/drm/i915/i915_gem_gtt.c 		pte |= BYT_PTE_SNOOPED_BY_CPU_CACHES;
pte               293 drivers/gpu/drm/i915/i915_gem_gtt.c 	return pte;
pte               300 drivers/gpu/drm/i915/i915_gem_gtt.c 	gen6_pte_t pte = GEN6_PTE_VALID;
pte               301 drivers/gpu/drm/i915/i915_gem_gtt.c 	pte |= HSW_PTE_ADDR_ENCODE(addr);
pte               304 drivers/gpu/drm/i915/i915_gem_gtt.c 		pte |= HSW_WB_LLC_AGE3;
pte               306 drivers/gpu/drm/i915/i915_gem_gtt.c 	return pte;
pte               313 drivers/gpu/drm/i915/i915_gem_gtt.c 	gen6_pte_t pte = GEN6_PTE_VALID;
pte               314 drivers/gpu/drm/i915/i915_gem_gtt.c 	pte |= HSW_PTE_ADDR_ENCODE(addr);
pte               320 drivers/gpu/drm/i915/i915_gem_gtt.c 		pte |= HSW_WT_ELLC_LLC_AGE3;
pte               323 drivers/gpu/drm/i915/i915_gem_gtt.c 		pte |= HSW_WB_ELLC_LLC_AGE3;
pte               327 drivers/gpu/drm/i915/i915_gem_gtt.c 	return pte;
pte              1611 drivers/gpu/drm/i915/i915_gem_gtt.c 	unsigned int pte = first_entry % GEN6_PTES;
pte              1617 drivers/gpu/drm/i915/i915_gem_gtt.c 		const unsigned int count = min(num_entries, GEN6_PTES - pte);
pte              1636 drivers/gpu/drm/i915/i915_gem_gtt.c 		memset32(vaddr + pte, scratch_pte, count);
pte              1639 drivers/gpu/drm/i915/i915_gem_gtt.c 		pte = 0;
pte              2169 drivers/gpu/drm/i915/i915_gem_gtt.c static void gen8_set_pte(void __iomem *addr, gen8_pte_t pte)
pte              2171 drivers/gpu/drm/i915/i915_gem_gtt.c 	writeq(pte, addr);
pte              2181 drivers/gpu/drm/i915/i915_gem_gtt.c 	gen8_pte_t __iomem *pte =
pte              2184 drivers/gpu/drm/i915/i915_gem_gtt.c 	gen8_set_pte(pte, gen8_pte_encode(addr, level, 0));
pte              2224 drivers/gpu/drm/i915/i915_gem_gtt.c 	gen6_pte_t __iomem *pte =
pte              2227 drivers/gpu/drm/i915/i915_gem_gtt.c 	iowrite32(vm->pte_encode(addr, level, flags), pte);
pte                38 drivers/gpu/drm/i915/i915_mm.c static int remap_pfn(pte_t *pte, unsigned long addr, void *data)
pte                43 drivers/gpu/drm/i915/i915_mm.c 	set_pte_at(r->mm, addr, pte, pte_mkspecial(pfn_pte(r->pfn, r->prot)));
pte               275 drivers/gpu/drm/lima/lima_vm.c 				u32 pte = *pt++;
pte               277 drivers/gpu/drm/lima/lima_vm.c 				if (pte)
pte               278 drivers/gpu/drm/lima/lima_vm.c 					printk(KERN_INFO "  pt %03x:%08x\n", k, pte);
pte                78 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 	u32 pte[NVKM_VMM_LEVELS_MAX];
pte               103 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 			buf += sprintf(buf,  "%05x:", it->pte[lvl]);
pte               146 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 	u32 pdei = it->pte[it->lvl + 1];
pte               211 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 		pgt->pte[lpti] -= pten;
pte               221 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 		if (pgt->pte[pteb] & NVKM_VMM_PTE_SPTES) {
pte               223 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 				if (!(pgt->pte[ptei] & NVKM_VMM_PTE_SPTES))
pte               235 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 		pgt->pte[ptei] &= ~NVKM_VMM_PTE_VALID;
pte               237 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 			if (pgt->pte[ptei] & NVKM_VMM_PTE_SPTES)
pte               239 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 			pgt->pte[ptei] &= ~NVKM_VMM_PTE_VALID;
pte               242 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 		if (pgt->pte[pteb] & NVKM_VMM_PTE_SPARSE) {
pte               310 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 		pgt->pte[lpti] += pten;
pte               320 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 		if (pgt->pte[pteb] & NVKM_VMM_PTE_VALID) {
pte               322 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 				if (!(pgt->pte[ptei] & NVKM_VMM_PTE_VALID))
pte               334 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 		pgt->pte[ptei] |= NVKM_VMM_PTE_VALID;
pte               336 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 			if (pgt->pte[ptei] & NVKM_VMM_PTE_VALID)
pte               338 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 			pgt->pte[ptei] |= NVKM_VMM_PTE_VALID;
pte               341 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 		if (pgt->pte[pteb] & NVKM_VMM_PTE_SPARSE) {
pte               389 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 		memset(&pgt->pte[ptei], NVKM_VMM_PTE_SPARSE, ptes);
pte               401 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 		memset(&pt->pte[ptei], 0x00, sizeof(pt->pte[0]) * ptes);
pte               448 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 			bool spte = pgt->pte[ptei] & NVKM_VMM_PTE_SPTES;
pte               450 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 				bool next = pgt->pte[ptei] & NVKM_VMM_PTE_SPTES;
pte               460 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 				memset(&pgt->pte[pteb], 0x00, ptes);
pte               464 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 					pgt->pte[pteb++] |= NVKM_VMM_PTE_VALID;
pte               519 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 		it.pte[it.lvl] = bits & ((1 << desc[it.lvl].bits) - 1);
pte               535 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 		const u32 ptei = it.pte[0];
pte               540 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 			const u32 pdei = it.pte[it.lvl];
pte               575 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 		it.pte[it.lvl] += ptes;
pte               578 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 			while (it.pte[it.lvl] == (1 << desc[it.lvl].bits)) {
pte               579 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 				it.pte[it.lvl++] = 0;
pte               580 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 				it.pte[it.lvl]++;
pte               592 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 	addr = it.pte[it.max--];
pte               595 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 		addr |= it.pte[it.max];
pte                50 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.h 	u8 pte[];
pte               830 drivers/gpu/drm/radeon/radeon_vm.c 		uint64_t pte;
pte               843 drivers/gpu/drm/radeon/radeon_vm.c 		pte = radeon_bo_gpu_offset(pt);
pte               844 drivers/gpu/drm/radeon/radeon_vm.c 		pte += (addr & mask) * 8;
pte               846 drivers/gpu/drm/radeon/radeon_vm.c 		if ((last_pte + 8 * count) != pte) {
pte               855 drivers/gpu/drm/radeon/radeon_vm.c 			last_pte = pte;
pte                98 drivers/gpu/drm/v3d/v3d_mmu.c 		u32 pte = page_prot | page_address;
pte               105 drivers/gpu/drm/v3d/v3d_mmu.c 			v3d->pt[page++] = pte + i;
pte               490 drivers/iommu/amd_iommu.c static u64 *first_pte_l7(u64 *pte, unsigned long *page_size,
pte               496 drivers/iommu/amd_iommu.c 	pg_size  = PTE_PAGE_SIZE(*pte);
pte               499 drivers/iommu/amd_iommu.c 	fpte     = (u64 *)(((unsigned long)pte) & pte_mask);
pte              1470 drivers/iommu/amd_iommu.c 	u64 *pte;
pte              1478 drivers/iommu/amd_iommu.c 	pte = (void *)get_zeroed_page(gfp);
pte              1479 drivers/iommu/amd_iommu.c 	if (!pte)
pte              1482 drivers/iommu/amd_iommu.c 	*pte             = PM_LEVEL_PDE(domain->mode,
pte              1484 drivers/iommu/amd_iommu.c 	domain->pt_root  = pte;
pte              1503 drivers/iommu/amd_iommu.c 	u64 *pte, *page;
pte              1511 drivers/iommu/amd_iommu.c 	pte     = &domain->pt_root[PM_LEVEL_INDEX(level, address)];
pte              1519 drivers/iommu/amd_iommu.c 		__pte     = *pte;
pte              1531 drivers/iommu/amd_iommu.c 			lpte = first_pte_l7(pte, NULL, &count);
pte              1554 drivers/iommu/amd_iommu.c 			if (cmpxchg64(pte, __pte, __npte) != __pte)
pte              1568 drivers/iommu/amd_iommu.c 		pte = IOMMU_PTE_PAGE(__pte);
pte              1571 drivers/iommu/amd_iommu.c 			*pte_page = pte;
pte              1573 drivers/iommu/amd_iommu.c 		pte = &pte[PM_LEVEL_INDEX(level, address)];
pte              1576 drivers/iommu/amd_iommu.c 	return pte;
pte              1588 drivers/iommu/amd_iommu.c 	u64 *pte;
pte              1596 drivers/iommu/amd_iommu.c 	pte	   = &domain->pt_root[PM_LEVEL_INDEX(level, address)];
pte              1602 drivers/iommu/amd_iommu.c 		if (!IOMMU_PTE_PRESENT(*pte))
pte              1606 drivers/iommu/amd_iommu.c 		if (PM_PTE_LEVEL(*pte) == 7 ||
pte              1607 drivers/iommu/amd_iommu.c 		    PM_PTE_LEVEL(*pte) == 0)
pte              1611 drivers/iommu/amd_iommu.c 		if (PM_PTE_LEVEL(*pte) != level)
pte              1617 drivers/iommu/amd_iommu.c 		pte	   = IOMMU_PTE_PAGE(*pte);
pte              1618 drivers/iommu/amd_iommu.c 		pte	   = &pte[PM_LEVEL_INDEX(level, address)];
pte              1626 drivers/iommu/amd_iommu.c 	if (PM_PTE_LEVEL(*pte) == PAGE_MODE_7_LEVEL)
pte              1627 drivers/iommu/amd_iommu.c 		pte = first_pte_l7(pte, page_size, NULL);
pte              1629 drivers/iommu/amd_iommu.c 	return pte;
pte              1632 drivers/iommu/amd_iommu.c static struct page *free_clear_pte(u64 *pte, u64 pteval, struct page *freelist)
pte              1637 drivers/iommu/amd_iommu.c 	while (cmpxchg64(pte, pteval, 0) != pteval) {
pte              1639 drivers/iommu/amd_iommu.c 		pteval = *pte;
pte              1667 drivers/iommu/amd_iommu.c 	u64 __pte, *pte;
pte              1678 drivers/iommu/amd_iommu.c 	pte   = alloc_pte(dom, bus_addr, page_size, NULL, gfp, &updated);
pte              1681 drivers/iommu/amd_iommu.c 	if (!pte)
pte              1685 drivers/iommu/amd_iommu.c 		freelist = free_clear_pte(&pte[i], pte[i], freelist);
pte              1702 drivers/iommu/amd_iommu.c 		pte[i] = __pte;
pte              1727 drivers/iommu/amd_iommu.c 	u64 *pte;
pte              1735 drivers/iommu/amd_iommu.c 		pte = fetch_pte(dom, bus_addr, &unmap_size);
pte              1737 drivers/iommu/amd_iommu.c 		if (pte) {
pte              1742 drivers/iommu/amd_iommu.c 				pte[i] = 0ULL;
pte              3144 drivers/iommu/amd_iommu.c 	u64 *pte, __pte;
pte              3149 drivers/iommu/amd_iommu.c 	pte = fetch_pte(domain, iova, &pte_pgsize);
pte              3151 drivers/iommu/amd_iommu.c 	if (!pte || !IOMMU_PTE_PRESENT(*pte))
pte              3155 drivers/iommu/amd_iommu.c 	__pte	    = __sme_clr(*pte & PM_ADDR_MASK);
pte              3489 drivers/iommu/amd_iommu.c 	u64 *pte;
pte              3494 drivers/iommu/amd_iommu.c 		pte   = &root[index];
pte              3499 drivers/iommu/amd_iommu.c 		if (!(*pte & GCR3_VALID)) {
pte              3507 drivers/iommu/amd_iommu.c 			*pte = iommu_virt_to_phys(root) | GCR3_VALID;
pte              3510 drivers/iommu/amd_iommu.c 		root = iommu_phys_to_virt(*pte & PAGE_MASK);
pte              3515 drivers/iommu/amd_iommu.c 	return pte;
pte              3521 drivers/iommu/amd_iommu.c 	u64 *pte;
pte              3526 drivers/iommu/amd_iommu.c 	pte = __get_gcr3_pte(domain->gcr3_tbl, domain->glx, pasid, true);
pte              3527 drivers/iommu/amd_iommu.c 	if (pte == NULL)
pte              3530 drivers/iommu/amd_iommu.c 	*pte = (cr3 & PAGE_MASK) | GCR3_VALID;
pte              3537 drivers/iommu/amd_iommu.c 	u64 *pte;
pte              3542 drivers/iommu/amd_iommu.c 	pte = __get_gcr3_pte(domain->gcr3_tbl, domain->glx, pasid, false);
pte              3543 drivers/iommu/amd_iommu.c 	if (pte == NULL)
pte              3546 drivers/iommu/amd_iommu.c 	*pte = 0;
pte               277 drivers/iommu/amd_iommu_types.h #define PM_PTE_LEVEL(pte)	(((pte) >> 9) & 0x7ULL)
pte               316 drivers/iommu/amd_iommu_types.h #define PTE_PAGE_SIZE(pte) \
pte               317 drivers/iommu/amd_iommu_types.h 	(1ULL << (1 + ffz(((pte) | 0xfffULL))))
pte               364 drivers/iommu/amd_iommu_types.h #define IOMMU_PTE_PRESENT(pte) ((pte) & IOMMU_PTE_PR)
pte               365 drivers/iommu/amd_iommu_types.h #define IOMMU_PTE_PAGE(pte) (iommu_phys_to_virt((pte) & IOMMU_PAGE_MASK))
pte               366 drivers/iommu/amd_iommu_types.h #define IOMMU_PTE_MODE(pte) (((pte) >> 9) & 0x07)
pte               884 drivers/iommu/intel-iommu.c 	struct dma_pte *parent, *pte;
pte               900 drivers/iommu/intel-iommu.c 		pte = &parent[offset];
pte               901 drivers/iommu/intel-iommu.c 		if (!*target_level && (dma_pte_superpage(pte) || !dma_pte_present(pte)))
pte               906 drivers/iommu/intel-iommu.c 		if (!dma_pte_present(pte)) {
pte               916 drivers/iommu/intel-iommu.c 			if (cmpxchg64(&pte->val, 0ULL, pteval))
pte               920 drivers/iommu/intel-iommu.c 				domain_flush_cache(domain, pte, sizeof(*pte));
pte               925 drivers/iommu/intel-iommu.c 		parent = phys_to_virt(dma_pte_addr(pte));
pte               932 drivers/iommu/intel-iommu.c 	return pte;
pte               940 drivers/iommu/intel-iommu.c 	struct dma_pte *parent, *pte;
pte               947 drivers/iommu/intel-iommu.c 		pte = &parent[offset];
pte               949 drivers/iommu/intel-iommu.c 			return pte;
pte               951 drivers/iommu/intel-iommu.c 		if (!dma_pte_present(pte)) {
pte               956 drivers/iommu/intel-iommu.c 		if (dma_pte_superpage(pte)) {
pte               958 drivers/iommu/intel-iommu.c 			return pte;
pte               961 drivers/iommu/intel-iommu.c 		parent = phys_to_virt(dma_pte_addr(pte));
pte               973 drivers/iommu/intel-iommu.c 	struct dma_pte *first_pte, *pte;
pte               982 drivers/iommu/intel-iommu.c 		first_pte = pte = dma_pfn_level_pte(domain, start_pfn, 1, &large_page);
pte               983 drivers/iommu/intel-iommu.c 		if (!pte) {
pte               988 drivers/iommu/intel-iommu.c 			dma_clear_pte(pte);
pte               990 drivers/iommu/intel-iommu.c 			pte++;
pte               991 drivers/iommu/intel-iommu.c 		} while (start_pfn <= last_pfn && !first_pte_in_page(pte));
pte               994 drivers/iommu/intel-iommu.c 				   (void *)pte - (void *)first_pte);
pte              1000 drivers/iommu/intel-iommu.c 			       int retain_level, struct dma_pte *pte,
pte              1005 drivers/iommu/intel-iommu.c 	pte = &pte[pfn_level_offset(pfn, level)];
pte              1011 drivers/iommu/intel-iommu.c 		if (!dma_pte_present(pte) || dma_pte_superpage(pte))
pte              1015 drivers/iommu/intel-iommu.c 		level_pte = phys_to_virt(dma_pte_addr(pte));
pte              1029 drivers/iommu/intel-iommu.c 			dma_clear_pte(pte);
pte              1030 drivers/iommu/intel-iommu.c 			domain_flush_cache(domain, pte, sizeof(*pte));
pte              1035 drivers/iommu/intel-iommu.c 	} while (!first_pte_in_page(++pte) && pfn <= last_pfn);
pte              1071 drivers/iommu/intel-iommu.c 					    int level, struct dma_pte *pte,
pte              1076 drivers/iommu/intel-iommu.c 	pg = pfn_to_page(dma_pte_addr(pte) >> PAGE_SHIFT);
pte              1083 drivers/iommu/intel-iommu.c 	pte = page_address(pg);
pte              1085 drivers/iommu/intel-iommu.c 		if (dma_pte_present(pte) && !dma_pte_superpage(pte))
pte              1087 drivers/iommu/intel-iommu.c 							   pte, freelist);
pte              1088 drivers/iommu/intel-iommu.c 		pte++;
pte              1089 drivers/iommu/intel-iommu.c 	} while (!first_pte_in_page(pte));
pte              1095 drivers/iommu/intel-iommu.c 					struct dma_pte *pte, unsigned long pfn,
pte              1103 drivers/iommu/intel-iommu.c 	pte = &pte[pfn_level_offset(pfn, level)];
pte              1108 drivers/iommu/intel-iommu.c 		if (!dma_pte_present(pte))
pte              1118 drivers/iommu/intel-iommu.c 			if (level > 1 && !dma_pte_superpage(pte))
pte              1119 drivers/iommu/intel-iommu.c 				freelist = dma_pte_list_pagetables(domain, level - 1, pte, freelist);
pte              1121 drivers/iommu/intel-iommu.c 			dma_clear_pte(pte);
pte              1123 drivers/iommu/intel-iommu.c 				first_pte = pte;
pte              1124 drivers/iommu/intel-iommu.c 			last_pte = pte;
pte              1128 drivers/iommu/intel-iommu.c 						       phys_to_virt(dma_pte_addr(pte)),
pte              1134 drivers/iommu/intel-iommu.c 	} while (!first_pte_in_page(++pte) && pfn <= last_pfn);
pte              2227 drivers/iommu/intel-iommu.c 	struct dma_pte *first_pte = NULL, *pte = NULL;
pte              2258 drivers/iommu/intel-iommu.c 		if (!pte) {
pte              2261 drivers/iommu/intel-iommu.c 			first_pte = pte = pfn_to_dma_pte(domain, iov_pfn, &largepage_lvl);
pte              2262 drivers/iommu/intel-iommu.c 			if (!pte)
pte              2290 drivers/iommu/intel-iommu.c 		tmp = cmpxchg64_local(&pte->val, 0ULL, pteval);
pte              2324 drivers/iommu/intel-iommu.c 		pte++;
pte              2325 drivers/iommu/intel-iommu.c 		if (!nr_pages || first_pte_in_page(pte) ||
pte              2328 drivers/iommu/intel-iommu.c 					   (void *)pte - (void *)first_pte);
pte              2329 drivers/iommu/intel-iommu.c 			pte = NULL;
pte              5368 drivers/iommu/intel-iommu.c 		struct dma_pte *pte;
pte              5370 drivers/iommu/intel-iommu.c 		pte = dmar_domain->pgd;
pte              5371 drivers/iommu/intel-iommu.c 		if (dma_pte_present(pte)) {
pte              5373 drivers/iommu/intel-iommu.c 				phys_to_virt(dma_pte_addr(pte));
pte              5374 drivers/iommu/intel-iommu.c 			free_pgtable_page(pte);
pte              5517 drivers/iommu/intel-iommu.c 	struct dma_pte *pte;
pte              5521 drivers/iommu/intel-iommu.c 	pte = pfn_to_dma_pte(dmar_domain, iova >> VTD_PAGE_SHIFT, &level);
pte              5522 drivers/iommu/intel-iommu.c 	if (pte && dma_pte_present(pte))
pte              5523 drivers/iommu/intel-iommu.c 		phys = dma_pte_addr(pte) +
pte               447 drivers/iommu/intel-pasid.c 	struct pasid_entry *pte;
pte               450 drivers/iommu/intel-pasid.c 	pte = intel_pasid_get_entry(dev, pasid);
pte               451 drivers/iommu/intel-pasid.c 	if (WARN_ON(!pte))
pte               454 drivers/iommu/intel-pasid.c 	did = pasid_get_domain_id(pte);
pte               458 drivers/iommu/intel-pasid.c 		clflush_cache_range(pte, sizeof(*pte));
pte               476 drivers/iommu/intel-pasid.c 	struct pasid_entry *pte;
pte               484 drivers/iommu/intel-pasid.c 	pte = intel_pasid_get_entry(dev, pasid);
pte               485 drivers/iommu/intel-pasid.c 	if (WARN_ON(!pte))
pte               488 drivers/iommu/intel-pasid.c 	pasid_clear_entry(pte);
pte               491 drivers/iommu/intel-pasid.c 	pasid_set_flptr(pte, (u64)__pa(pgd));
pte               498 drivers/iommu/intel-pasid.c 		pasid_set_sre(pte);
pte               505 drivers/iommu/intel-pasid.c 			pasid_set_flpm(pte, 1);
pte               508 drivers/iommu/intel-pasid.c 			pasid_clear_entry(pte);
pte               514 drivers/iommu/intel-pasid.c 	pasid_set_domain_id(pte, did);
pte               515 drivers/iommu/intel-pasid.c 	pasid_set_address_width(pte, iommu->agaw);
pte               516 drivers/iommu/intel-pasid.c 	pasid_set_page_snoop(pte, !!ecap_smpwc(iommu->ecap));
pte               519 drivers/iommu/intel-pasid.c 	pasid_set_translation_type(pte, 1);
pte               520 drivers/iommu/intel-pasid.c 	pasid_set_present(pte);
pte               523 drivers/iommu/intel-pasid.c 		clflush_cache_range(pte, sizeof(*pte));
pte               542 drivers/iommu/intel-pasid.c 	struct pasid_entry *pte;
pte               574 drivers/iommu/intel-pasid.c 	pte = intel_pasid_get_entry(dev, pasid);
pte               575 drivers/iommu/intel-pasid.c 	if (!pte) {
pte               580 drivers/iommu/intel-pasid.c 	pasid_clear_entry(pte);
pte               581 drivers/iommu/intel-pasid.c 	pasid_set_domain_id(pte, did);
pte               582 drivers/iommu/intel-pasid.c 	pasid_set_slptr(pte, pgd_val);
pte               583 drivers/iommu/intel-pasid.c 	pasid_set_address_width(pte, agaw);
pte               584 drivers/iommu/intel-pasid.c 	pasid_set_translation_type(pte, 2);
pte               585 drivers/iommu/intel-pasid.c 	pasid_set_fault_enable(pte);
pte               586 drivers/iommu/intel-pasid.c 	pasid_set_page_snoop(pte, !!ecap_smpwc(iommu->ecap));
pte               592 drivers/iommu/intel-pasid.c 	pasid_set_sre(pte);
pte               593 drivers/iommu/intel-pasid.c 	pasid_set_present(pte);
pte               596 drivers/iommu/intel-pasid.c 		clflush_cache_range(pte, sizeof(*pte));
pte               616 drivers/iommu/intel-pasid.c 	struct pasid_entry *pte;
pte               618 drivers/iommu/intel-pasid.c 	pte = intel_pasid_get_entry(dev, pasid);
pte               619 drivers/iommu/intel-pasid.c 	if (!pte) {
pte               624 drivers/iommu/intel-pasid.c 	pasid_clear_entry(pte);
pte               625 drivers/iommu/intel-pasid.c 	pasid_set_domain_id(pte, did);
pte               626 drivers/iommu/intel-pasid.c 	pasid_set_address_width(pte, iommu->agaw);
pte               627 drivers/iommu/intel-pasid.c 	pasid_set_translation_type(pte, 4);
pte               628 drivers/iommu/intel-pasid.c 	pasid_set_fault_enable(pte);
pte               629 drivers/iommu/intel-pasid.c 	pasid_set_page_snoop(pte, !!ecap_smpwc(iommu->ecap));
pte               635 drivers/iommu/intel-pasid.c 	pasid_set_sre(pte);
pte               636 drivers/iommu/intel-pasid.c 	pasid_set_present(pte);
pte               639 drivers/iommu/intel-pasid.c 		clflush_cache_range(pte, sizeof(*pte));
pte                73 drivers/iommu/intel-pasid.h static inline bool pasid_pte_is_present(struct pasid_entry *pte)
pte                75 drivers/iommu/intel-pasid.h 	return READ_ONCE(pte->val[0]) & PASID_PTE_PRESENT;
pte                84 drivers/iommu/io-pgtable-arm-v7s.c #define ARM_V7S_PTE_IS_VALID(pte)	(((pte) & 0x3) != 0)
pte                85 drivers/iommu/io-pgtable-arm-v7s.c #define ARM_V7S_PTE_IS_TABLE(pte, lvl) \
pte                86 drivers/iommu/io-pgtable-arm-v7s.c 	((lvl) == 1 && (((pte) & 0x3) == ARM_V7S_PTE_TYPE_TABLE))
pte               174 drivers/iommu/io-pgtable-arm-v7s.c static bool arm_v7s_pte_is_cont(arm_v7s_iopte pte, int lvl);
pte               190 drivers/iommu/io-pgtable-arm-v7s.c 	arm_v7s_iopte pte = paddr & ARM_V7S_LVL_MASK(lvl);
pte               193 drivers/iommu/io-pgtable-arm-v7s.c 		return pte;
pte               196 drivers/iommu/io-pgtable-arm-v7s.c 		pte |= ARM_V7S_ATTR_MTK_PA_BIT32;
pte               198 drivers/iommu/io-pgtable-arm-v7s.c 		pte |= ARM_V7S_ATTR_MTK_PA_BIT33;
pte               199 drivers/iommu/io-pgtable-arm-v7s.c 	return pte;
pte               202 drivers/iommu/io-pgtable-arm-v7s.c static phys_addr_t iopte_to_paddr(arm_v7s_iopte pte, int lvl,
pte               208 drivers/iommu/io-pgtable-arm-v7s.c 	if (ARM_V7S_PTE_IS_TABLE(pte, lvl))
pte               210 drivers/iommu/io-pgtable-arm-v7s.c 	else if (arm_v7s_pte_is_cont(pte, lvl))
pte               215 drivers/iommu/io-pgtable-arm-v7s.c 	paddr = pte & mask;
pte               219 drivers/iommu/io-pgtable-arm-v7s.c 	if (pte & ARM_V7S_ATTR_MTK_PA_BIT32)
pte               221 drivers/iommu/io-pgtable-arm-v7s.c 	if (pte & ARM_V7S_ATTR_MTK_PA_BIT33)
pte               226 drivers/iommu/io-pgtable-arm-v7s.c static arm_v7s_iopte *iopte_deref(arm_v7s_iopte pte, int lvl,
pte               229 drivers/iommu/io-pgtable-arm-v7s.c 	return phys_to_virt(iopte_to_paddr(pte, lvl, &data->iop.cfg));
pte               305 drivers/iommu/io-pgtable-arm-v7s.c static void __arm_v7s_set_pte(arm_v7s_iopte *ptep, arm_v7s_iopte pte,
pte               311 drivers/iommu/io-pgtable-arm-v7s.c 		ptep[i] = pte;
pte               320 drivers/iommu/io-pgtable-arm-v7s.c 	arm_v7s_iopte pte = ARM_V7S_ATTR_NG | ARM_V7S_ATTR_S;
pte               323 drivers/iommu/io-pgtable-arm-v7s.c 		pte |= ARM_V7S_ATTR_TEX(1);
pte               325 drivers/iommu/io-pgtable-arm-v7s.c 		pte |= ARM_V7S_PTE_AF;
pte               327 drivers/iommu/io-pgtable-arm-v7s.c 			pte |= ARM_V7S_PTE_AP_UNPRIV;
pte               329 drivers/iommu/io-pgtable-arm-v7s.c 			pte |= ARM_V7S_PTE_AP_RDONLY;
pte               331 drivers/iommu/io-pgtable-arm-v7s.c 	pte <<= ARM_V7S_ATTR_SHIFT(lvl);
pte               334 drivers/iommu/io-pgtable-arm-v7s.c 		pte |= ARM_V7S_ATTR_XN(lvl);
pte               336 drivers/iommu/io-pgtable-arm-v7s.c 		pte |= ARM_V7S_ATTR_B;
pte               338 drivers/iommu/io-pgtable-arm-v7s.c 		pte |= ARM_V7S_ATTR_B | ARM_V7S_ATTR_C;
pte               340 drivers/iommu/io-pgtable-arm-v7s.c 	pte |= ARM_V7S_PTE_TYPE_PAGE;
pte               342 drivers/iommu/io-pgtable-arm-v7s.c 		pte |= ARM_V7S_ATTR_NS_SECTION;
pte               344 drivers/iommu/io-pgtable-arm-v7s.c 	return pte;
pte               347 drivers/iommu/io-pgtable-arm-v7s.c static int arm_v7s_pte_to_prot(arm_v7s_iopte pte, int lvl)
pte               350 drivers/iommu/io-pgtable-arm-v7s.c 	arm_v7s_iopte attr = pte >> ARM_V7S_ATTR_SHIFT(lvl);
pte               358 drivers/iommu/io-pgtable-arm-v7s.c 	else if (pte & ARM_V7S_ATTR_C)
pte               360 drivers/iommu/io-pgtable-arm-v7s.c 	if (pte & ARM_V7S_ATTR_XN(lvl))
pte               366 drivers/iommu/io-pgtable-arm-v7s.c static arm_v7s_iopte arm_v7s_pte_to_cont(arm_v7s_iopte pte, int lvl)
pte               369 drivers/iommu/io-pgtable-arm-v7s.c 		pte |= ARM_V7S_CONT_SECTION;
pte               371 drivers/iommu/io-pgtable-arm-v7s.c 		arm_v7s_iopte xn = pte & ARM_V7S_ATTR_XN(lvl);
pte               372 drivers/iommu/io-pgtable-arm-v7s.c 		arm_v7s_iopte tex = pte & ARM_V7S_CONT_PAGE_TEX_MASK;
pte               374 drivers/iommu/io-pgtable-arm-v7s.c 		pte ^= xn | tex | ARM_V7S_PTE_TYPE_PAGE;
pte               375 drivers/iommu/io-pgtable-arm-v7s.c 		pte |= (xn << ARM_V7S_CONT_PAGE_XN_SHIFT) |
pte               379 drivers/iommu/io-pgtable-arm-v7s.c 	return pte;
pte               382 drivers/iommu/io-pgtable-arm-v7s.c static arm_v7s_iopte arm_v7s_cont_to_pte(arm_v7s_iopte pte, int lvl)
pte               385 drivers/iommu/io-pgtable-arm-v7s.c 		pte &= ~ARM_V7S_CONT_SECTION;
pte               387 drivers/iommu/io-pgtable-arm-v7s.c 		arm_v7s_iopte xn = pte & BIT(ARM_V7S_CONT_PAGE_XN_SHIFT);
pte               388 drivers/iommu/io-pgtable-arm-v7s.c 		arm_v7s_iopte tex = pte & (ARM_V7S_CONT_PAGE_TEX_MASK <<
pte               391 drivers/iommu/io-pgtable-arm-v7s.c 		pte ^= xn | tex | ARM_V7S_PTE_TYPE_CONT_PAGE;
pte               392 drivers/iommu/io-pgtable-arm-v7s.c 		pte |= (xn >> ARM_V7S_CONT_PAGE_XN_SHIFT) |
pte               396 drivers/iommu/io-pgtable-arm-v7s.c 	return pte;
pte               399 drivers/iommu/io-pgtable-arm-v7s.c static bool arm_v7s_pte_is_cont(arm_v7s_iopte pte, int lvl)
pte               401 drivers/iommu/io-pgtable-arm-v7s.c 	if (lvl == 1 && !ARM_V7S_PTE_IS_TABLE(pte, lvl))
pte               402 drivers/iommu/io-pgtable-arm-v7s.c 		return pte & ARM_V7S_CONT_SECTION;
pte               404 drivers/iommu/io-pgtable-arm-v7s.c 		return !(pte & ARM_V7S_PTE_TYPE_PAGE);
pte               417 drivers/iommu/io-pgtable-arm-v7s.c 	arm_v7s_iopte pte;
pte               439 drivers/iommu/io-pgtable-arm-v7s.c 	pte = arm_v7s_prot_to_pte(prot, lvl, cfg);
pte               441 drivers/iommu/io-pgtable-arm-v7s.c 		pte = arm_v7s_pte_to_cont(pte, lvl);
pte               443 drivers/iommu/io-pgtable-arm-v7s.c 	pte |= paddr_to_iopte(paddr, lvl, cfg);
pte               445 drivers/iommu/io-pgtable-arm-v7s.c 	__arm_v7s_set_pte(ptep, pte, num_entries, cfg);
pte               478 drivers/iommu/io-pgtable-arm-v7s.c 	arm_v7s_iopte pte, *cptep;
pte               494 drivers/iommu/io-pgtable-arm-v7s.c 	pte = READ_ONCE(*ptep);
pte               495 drivers/iommu/io-pgtable-arm-v7s.c 	if (!pte) {
pte               500 drivers/iommu/io-pgtable-arm-v7s.c 		pte = arm_v7s_install_table(cptep, ptep, 0, cfg);
pte               501 drivers/iommu/io-pgtable-arm-v7s.c 		if (pte)
pte               508 drivers/iommu/io-pgtable-arm-v7s.c 	if (ARM_V7S_PTE_IS_TABLE(pte, lvl)) {
pte               509 drivers/iommu/io-pgtable-arm-v7s.c 		cptep = iopte_deref(pte, lvl, data);
pte               510 drivers/iommu/io-pgtable-arm-v7s.c 	} else if (pte) {
pte               556 drivers/iommu/io-pgtable-arm-v7s.c 		arm_v7s_iopte pte = data->pgd[i];
pte               558 drivers/iommu/io-pgtable-arm-v7s.c 		if (ARM_V7S_PTE_IS_TABLE(pte, 1))
pte               559 drivers/iommu/io-pgtable-arm-v7s.c 			__arm_v7s_free_table(iopte_deref(pte, 1, data),
pte               572 drivers/iommu/io-pgtable-arm-v7s.c 	arm_v7s_iopte pte;
pte               577 drivers/iommu/io-pgtable-arm-v7s.c 	pte = *ptep;
pte               578 drivers/iommu/io-pgtable-arm-v7s.c 	if (!arm_v7s_pte_is_cont(pte, lvl))
pte               579 drivers/iommu/io-pgtable-arm-v7s.c 		return pte;
pte               582 drivers/iommu/io-pgtable-arm-v7s.c 	pte = arm_v7s_cont_to_pte(pte, lvl);
pte               584 drivers/iommu/io-pgtable-arm-v7s.c 		ptep[i] = pte + i * size;
pte               590 drivers/iommu/io-pgtable-arm-v7s.c 	return pte;
pte               600 drivers/iommu/io-pgtable-arm-v7s.c 	arm_v7s_iopte pte, *tablep;
pte               611 drivers/iommu/io-pgtable-arm-v7s.c 	pte = arm_v7s_prot_to_pte(arm_v7s_pte_to_prot(blk_pte, 1), 2, cfg);
pte               613 drivers/iommu/io-pgtable-arm-v7s.c 		pte = arm_v7s_pte_to_cont(pte, 2);
pte               615 drivers/iommu/io-pgtable-arm-v7s.c 	for (i = 0; i < num_ptes; i += num_entries, pte += size) {
pte               620 drivers/iommu/io-pgtable-arm-v7s.c 		__arm_v7s_set_pte(&tablep[i], pte, num_entries, cfg);
pte               623 drivers/iommu/io-pgtable-arm-v7s.c 	pte = arm_v7s_install_table(tablep, ptep, blk_pte, cfg);
pte               624 drivers/iommu/io-pgtable-arm-v7s.c 	if (pte != blk_pte) {
pte               627 drivers/iommu/io-pgtable-arm-v7s.c 		if (!ARM_V7S_PTE_IS_TABLE(pte, 1))
pte               630 drivers/iommu/io-pgtable-arm-v7s.c 		tablep = iopte_deref(pte, 1, data);
pte               643 drivers/iommu/io-pgtable-arm-v7s.c 	arm_v7s_iopte pte[ARM_V7S_CONT_PAGES];
pte               654 drivers/iommu/io-pgtable-arm-v7s.c 		pte[i] = READ_ONCE(ptep[i]);
pte               655 drivers/iommu/io-pgtable-arm-v7s.c 		if (WARN_ON(!ARM_V7S_PTE_IS_VALID(pte[i])))
pte               669 drivers/iommu/io-pgtable-arm-v7s.c 	if (num_entries <= 1 && arm_v7s_pte_is_cont(pte[0], lvl)) {
pte               673 drivers/iommu/io-pgtable-arm-v7s.c 		pte[0] = arm_v7s_split_cont(data, iova, idx, lvl, ptep);
pte               684 drivers/iommu/io-pgtable-arm-v7s.c 			if (ARM_V7S_PTE_IS_TABLE(pte[i], lvl)) {
pte               688 drivers/iommu/io-pgtable-arm-v7s.c 				ptep = iopte_deref(pte[i], lvl, data);
pte               703 drivers/iommu/io-pgtable-arm-v7s.c 	} else if (lvl == 1 && !ARM_V7S_PTE_IS_TABLE(pte[0], lvl)) {
pte               708 drivers/iommu/io-pgtable-arm-v7s.c 		return arm_v7s_split_blk_unmap(data, gather, iova, size, pte[0],
pte               713 drivers/iommu/io-pgtable-arm-v7s.c 	ptep = iopte_deref(pte[0], lvl, data);
pte               732 drivers/iommu/io-pgtable-arm-v7s.c 	arm_v7s_iopte *ptep = data->pgd, pte;
pte               738 drivers/iommu/io-pgtable-arm-v7s.c 		pte = READ_ONCE(*ptep);
pte               739 drivers/iommu/io-pgtable-arm-v7s.c 		ptep = iopte_deref(pte, lvl, data);
pte               740 drivers/iommu/io-pgtable-arm-v7s.c 	} while (ARM_V7S_PTE_IS_TABLE(pte, lvl));
pte               742 drivers/iommu/io-pgtable-arm-v7s.c 	if (!ARM_V7S_PTE_IS_VALID(pte))
pte               746 drivers/iommu/io-pgtable-arm-v7s.c 	if (arm_v7s_pte_is_cont(pte, lvl))
pte               748 drivers/iommu/io-pgtable-arm-v7s.c 	return iopte_to_paddr(pte, lvl, &data->iop.cfg) | (iova & ~mask);
pte               173 drivers/iommu/io-pgtable-arm.c #define iopte_deref(pte,d) __va(iopte_to_paddr(pte, d))
pte               175 drivers/iommu/io-pgtable-arm.c #define iopte_type(pte,l)					\
pte               176 drivers/iommu/io-pgtable-arm.c 	(((pte) >> ARM_LPAE_PTE_TYPE_SHIFT) & ARM_LPAE_PTE_TYPE_MASK)
pte               178 drivers/iommu/io-pgtable-arm.c #define iopte_prot(pte)	((pte) & ARM_LPAE_PTE_ATTR_MASK)
pte               193 drivers/iommu/io-pgtable-arm.c static inline bool iopte_leaf(arm_lpae_iopte pte, int lvl,
pte               197 drivers/iommu/io-pgtable-arm.c 		return iopte_type(pte, lvl) == ARM_LPAE_PTE_TYPE_PAGE;
pte               199 drivers/iommu/io-pgtable-arm.c 	return iopte_type(pte, lvl) == ARM_LPAE_PTE_TYPE_BLOCK;
pte               205 drivers/iommu/io-pgtable-arm.c 	arm_lpae_iopte pte = paddr;
pte               208 drivers/iommu/io-pgtable-arm.c 	return (pte | (pte >> (48 - 12))) & ARM_LPAE_PTE_ADDR_MASK;
pte               211 drivers/iommu/io-pgtable-arm.c static phys_addr_t iopte_to_paddr(arm_lpae_iopte pte,
pte               214 drivers/iommu/io-pgtable-arm.c 	u64 paddr = pte & ARM_LPAE_PTE_ADDR_MASK;
pte               285 drivers/iommu/io-pgtable-arm.c static void __arm_lpae_set_pte(arm_lpae_iopte *ptep, arm_lpae_iopte pte,
pte               288 drivers/iommu/io-pgtable-arm.c 	*ptep = pte;
pte               303 drivers/iommu/io-pgtable-arm.c 	arm_lpae_iopte pte = prot;
pte               306 drivers/iommu/io-pgtable-arm.c 		pte |= ARM_LPAE_PTE_NS;
pte               309 drivers/iommu/io-pgtable-arm.c 		pte |= ARM_LPAE_PTE_TYPE_PAGE;
pte               311 drivers/iommu/io-pgtable-arm.c 		pte |= ARM_LPAE_PTE_TYPE_BLOCK;
pte               314 drivers/iommu/io-pgtable-arm.c 		pte |= ARM_LPAE_PTE_AF;
pte               315 drivers/iommu/io-pgtable-arm.c 	pte |= ARM_LPAE_PTE_SH_IS;
pte               316 drivers/iommu/io-pgtable-arm.c 	pte |= paddr_to_iopte(paddr, data);
pte               318 drivers/iommu/io-pgtable-arm.c 	__arm_lpae_set_pte(ptep, pte, &data->iop.cfg);
pte               326 drivers/iommu/io-pgtable-arm.c 	arm_lpae_iopte pte = *ptep;
pte               328 drivers/iommu/io-pgtable-arm.c 	if (iopte_leaf(pte, lvl, data->iop.fmt)) {
pte               332 drivers/iommu/io-pgtable-arm.c 	} else if (iopte_type(pte, lvl) == ARM_LPAE_PTE_TYPE_TABLE) {
pte               386 drivers/iommu/io-pgtable-arm.c 	arm_lpae_iopte *cptep, pte;
pte               403 drivers/iommu/io-pgtable-arm.c 	pte = READ_ONCE(*ptep);
pte               404 drivers/iommu/io-pgtable-arm.c 	if (!pte) {
pte               409 drivers/iommu/io-pgtable-arm.c 		pte = arm_lpae_install_table(cptep, ptep, 0, cfg);
pte               410 drivers/iommu/io-pgtable-arm.c 		if (pte)
pte               412 drivers/iommu/io-pgtable-arm.c 	} else if (!cfg->coherent_walk && !(pte & ARM_LPAE_PTE_SW_SYNC)) {
pte               416 drivers/iommu/io-pgtable-arm.c 	if (pte && !iopte_leaf(pte, lvl, data->iop.fmt)) {
pte               417 drivers/iommu/io-pgtable-arm.c 		cptep = iopte_deref(pte, data);
pte               418 drivers/iommu/io-pgtable-arm.c 	} else if (pte) {
pte               431 drivers/iommu/io-pgtable-arm.c 	arm_lpae_iopte pte;
pte               435 drivers/iommu/io-pgtable-arm.c 		pte = ARM_LPAE_PTE_nG;
pte               437 drivers/iommu/io-pgtable-arm.c 			pte |= ARM_LPAE_PTE_AP_RDONLY;
pte               439 drivers/iommu/io-pgtable-arm.c 			pte |= ARM_LPAE_PTE_AP_UNPRIV;
pte               441 drivers/iommu/io-pgtable-arm.c 		pte = ARM_LPAE_PTE_HAP_FAULT;
pte               443 drivers/iommu/io-pgtable-arm.c 			pte |= ARM_LPAE_PTE_HAP_READ;
pte               445 drivers/iommu/io-pgtable-arm.c 			pte |= ARM_LPAE_PTE_HAP_WRITE;
pte               455 drivers/iommu/io-pgtable-arm.c 			pte |= ARM_LPAE_PTE_MEMATTR_DEV;
pte               457 drivers/iommu/io-pgtable-arm.c 			pte |= ARM_LPAE_PTE_MEMATTR_OIWB;
pte               459 drivers/iommu/io-pgtable-arm.c 			pte |= ARM_LPAE_PTE_MEMATTR_NC;
pte               462 drivers/iommu/io-pgtable-arm.c 			pte |= (ARM_LPAE_MAIR_ATTR_IDX_DEV
pte               465 drivers/iommu/io-pgtable-arm.c 			pte |= (ARM_LPAE_MAIR_ATTR_IDX_CACHE
pte               468 drivers/iommu/io-pgtable-arm.c 			pte |= (ARM_LPAE_MAIR_ATTR_IDX_INC_OCACHE
pte               473 drivers/iommu/io-pgtable-arm.c 		pte |= ARM_LPAE_PTE_XN;
pte               475 drivers/iommu/io-pgtable-arm.c 	return pte;
pte               525 drivers/iommu/io-pgtable-arm.c 		arm_lpae_iopte pte = *ptep++;
pte               527 drivers/iommu/io-pgtable-arm.c 		if (!pte || iopte_leaf(pte, lvl, data->iop.fmt))
pte               530 drivers/iommu/io-pgtable-arm.c 		__arm_lpae_free_pgtable(data, lvl + 1, iopte_deref(pte, data));
pte               551 drivers/iommu/io-pgtable-arm.c 	arm_lpae_iopte pte, *tablep;
pte               568 drivers/iommu/io-pgtable-arm.c 	pte = iopte_prot(blk_pte);
pte               570 drivers/iommu/io-pgtable-arm.c 	for (i = 0; i < tablesz / sizeof(pte); i++, blk_paddr += split_sz) {
pte               575 drivers/iommu/io-pgtable-arm.c 		__arm_lpae_init_pte(data, blk_paddr, pte, lvl, &tablep[i]);
pte               578 drivers/iommu/io-pgtable-arm.c 	pte = arm_lpae_install_table(tablep, ptep, blk_pte, cfg);
pte               579 drivers/iommu/io-pgtable-arm.c 	if (pte != blk_pte) {
pte               586 drivers/iommu/io-pgtable-arm.c 		if (iopte_type(pte, lvl - 1) != ARM_LPAE_PTE_TYPE_TABLE)
pte               589 drivers/iommu/io-pgtable-arm.c 		tablep = iopte_deref(pte, data);
pte               603 drivers/iommu/io-pgtable-arm.c 	arm_lpae_iopte pte;
pte               611 drivers/iommu/io-pgtable-arm.c 	pte = READ_ONCE(*ptep);
pte               612 drivers/iommu/io-pgtable-arm.c 	if (WARN_ON(!pte))
pte               619 drivers/iommu/io-pgtable-arm.c 		if (!iopte_leaf(pte, lvl, iop->fmt)) {
pte               623 drivers/iommu/io-pgtable-arm.c 			ptep = iopte_deref(pte, data);
pte               637 drivers/iommu/io-pgtable-arm.c 	} else if (iopte_leaf(pte, lvl, iop->fmt)) {
pte               642 drivers/iommu/io-pgtable-arm.c 		return arm_lpae_split_blk_unmap(data, gather, iova, size, pte,
pte               647 drivers/iommu/io-pgtable-arm.c 	ptep = iopte_deref(pte, data);
pte               668 drivers/iommu/io-pgtable-arm.c 	arm_lpae_iopte pte, *ptep = data->pgd;
pte               678 drivers/iommu/io-pgtable-arm.c 		pte = READ_ONCE(*ptep);
pte               681 drivers/iommu/io-pgtable-arm.c 		if (!pte)
pte               685 drivers/iommu/io-pgtable-arm.c 		if (iopte_leaf(pte, lvl, data->iop.fmt))
pte               689 drivers/iommu/io-pgtable-arm.c 		ptep = iopte_deref(pte, data);
pte               697 drivers/iommu/io-pgtable-arm.c 	return iopte_to_paddr(pte, data) | iova;
pte              1630 drivers/iommu/omap-iommu.c 	u32 *pgd, *pte;
pte              1637 drivers/iommu/omap-iommu.c 	iopgtable_lookup_entry(oiommu, da, &pgd, &pte);
pte              1639 drivers/iommu/omap-iommu.c 	if (pte) {
pte              1640 drivers/iommu/omap-iommu.c 		if (iopte_is_small(*pte))
pte              1641 drivers/iommu/omap-iommu.c 			ret = omap_iommu_translate(*pte, da, IOPTE_MASK);
pte              1642 drivers/iommu/omap-iommu.c 		else if (iopte_is_large(*pte))
pte              1643 drivers/iommu/omap-iommu.c 			ret = omap_iommu_translate(*pte, da, IOLARGE_MASK);
pte              1645 drivers/iommu/omap-iommu.c 			dev_err(dev, "bogus pte 0x%x, da 0x%llx", *pte,
pte               218 drivers/iommu/rockchip-iommu.c static inline phys_addr_t rk_pte_page_address(u32 pte)
pte               220 drivers/iommu/rockchip-iommu.c 	return (phys_addr_t)pte & RK_PTE_PAGE_ADDRESS_MASK;
pte               223 drivers/iommu/rockchip-iommu.c static inline bool rk_pte_is_page_valid(u32 pte)
pte               225 drivers/iommu/rockchip-iommu.c 	return pte & RK_PTE_PAGE_VALID;
pte               238 drivers/iommu/rockchip-iommu.c static u32 rk_mk_pte_invalid(u32 pte)
pte               240 drivers/iommu/rockchip-iommu.c 	return pte & ~RK_PTE_PAGE_VALID;
pte               483 drivers/iommu/rockchip-iommu.c 	u32 pte = 0;
pte               503 drivers/iommu/rockchip-iommu.c 	pte = *pte_addr;
pte               505 drivers/iommu/rockchip-iommu.c 	if (!rk_pte_is_page_valid(pte))
pte               508 drivers/iommu/rockchip-iommu.c 	page_addr_phys = rk_pte_page_address(pte) + page_offset;
pte               509 drivers/iommu/rockchip-iommu.c 	page_flags = pte & RK_PTE_PAGE_FLAGS_MASK;
pte               516 drivers/iommu/rockchip-iommu.c 		rk_dte_is_pt_valid(dte), &pte_addr_phys, pte,
pte               517 drivers/iommu/rockchip-iommu.c 		rk_pte_is_page_valid(pte), &page_addr_phys, page_flags);
pte               595 drivers/iommu/rockchip-iommu.c 	u32 dte, pte;
pte               606 drivers/iommu/rockchip-iommu.c 	pte = page_table[rk_iova_pte_index(iova)];
pte               607 drivers/iommu/rockchip-iommu.c 	if (!rk_pte_is_page_valid(pte))
pte               610 drivers/iommu/rockchip-iommu.c 	phys = rk_pte_page_address(pte) + rk_iova_page_offset(iova);
pte               703 drivers/iommu/rockchip-iommu.c 		u32 pte = pte_addr[pte_count];
pte               704 drivers/iommu/rockchip-iommu.c 		if (!rk_pte_is_page_valid(pte))
pte               707 drivers/iommu/rockchip-iommu.c 		pte_addr[pte_count] = rk_mk_pte_invalid(pte);
pte               726 drivers/iommu/rockchip-iommu.c 		u32 pte = pte_addr[pte_count];
pte               728 drivers/iommu/rockchip-iommu.c 		if (rk_pte_is_page_valid(pte))
pte                66 drivers/iommu/tegra-gart.c 				unsigned long iova, unsigned long pte)
pte                69 drivers/iommu/tegra-gart.c 	writel_relaxed(pte, gart->regs + GART_ENTRY_DATA);
pte                75 drivers/iommu/tegra-gart.c 	unsigned long pte;
pte                78 drivers/iommu/tegra-gart.c 	pte = readl_relaxed(gart->regs + GART_ENTRY_DATA);
pte                80 drivers/iommu/tegra-gart.c 	return pte;
pte               229 drivers/iommu/tegra-gart.c 	unsigned long pte;
pte               235 drivers/iommu/tegra-gart.c 	pte = gart_read_pte(gart, iova);
pte               238 drivers/iommu/tegra-gart.c 	return pte & GART_PAGE_MASK;
pte               639 drivers/iommu/tegra-smmu.c 			       u32 *pte, dma_addr_t pte_dma, u32 val)
pte               642 drivers/iommu/tegra-smmu.c 	unsigned long offset = offset_in_page(pte);
pte               644 drivers/iommu/tegra-smmu.c 	*pte = val;
pte               659 drivers/iommu/tegra-smmu.c 	u32 *pte;
pte               661 drivers/iommu/tegra-smmu.c 	pte = as_get_pte(as, iova, &pte_dma);
pte               662 drivers/iommu/tegra-smmu.c 	if (!pte)
pte               666 drivers/iommu/tegra-smmu.c 	if (*pte == 0)
pte               677 drivers/iommu/tegra-smmu.c 	tegra_smmu_set_pte(as, iova, pte, pte_dma,
pte               688 drivers/iommu/tegra-smmu.c 	u32 *pte;
pte               690 drivers/iommu/tegra-smmu.c 	pte = tegra_smmu_pte_lookup(as, iova, &pte_dma);
pte               691 drivers/iommu/tegra-smmu.c 	if (!pte || !*pte)
pte               694 drivers/iommu/tegra-smmu.c 	tegra_smmu_set_pte(as, iova, pte, pte_dma, 0);
pte               706 drivers/iommu/tegra-smmu.c 	u32 *pte;
pte               708 drivers/iommu/tegra-smmu.c 	pte = tegra_smmu_pte_lookup(as, iova, &pte_dma);
pte               709 drivers/iommu/tegra-smmu.c 	if (!pte || !*pte)
pte               712 drivers/iommu/tegra-smmu.c 	pfn = *pte & as->smmu->pfn_mask;
pte               179 drivers/md/dm-switch.c 	region_table_slot_t pte;
pte               183 drivers/md/dm-switch.c 	pte = sctx->region_table[region_index];
pte               184 drivers/md/dm-switch.c 	pte &= ~((((region_table_slot_t)1 << sctx->region_table_entry_bits) - 1) << bit);
pte               185 drivers/md/dm-switch.c 	pte |= (region_table_slot_t)value << bit;
pte               186 drivers/md/dm-switch.c 	sctx->region_table[region_index] = pte;
pte               212 drivers/misc/sgi-gru/grufault.c 	pte_t pte;
pte               231 drivers/misc/sgi-gru/grufault.c 		pte = *(pte_t *) pmdp;
pte               234 drivers/misc/sgi-gru/grufault.c 		pte = *pte_offset_kernel(pmdp, vaddr);
pte               236 drivers/misc/sgi-gru/grufault.c 	if (unlikely(!pte_present(pte) ||
pte               237 drivers/misc/sgi-gru/grufault.c 		     (write && (!pte_write(pte) || !pte_dirty(pte)))))
pte               240 drivers/misc/sgi-gru/grufault.c 	*paddr = pte_pfn(pte) << PAGE_SHIFT;
pte               878 drivers/net/ethernet/mellanox/mlxsw/reg.h MLXSW_ITEM32(reg, spvm, pte, 0x00, 30, 1);
pte               320 drivers/staging/gasket/gasket_page_table.c 					  struct gasket_page_table_entry *pte,
pte               324 drivers/staging/gasket/gasket_page_table.c 	pte->status = PTE_FREE;
pte               329 drivers/staging/gasket/gasket_page_table.c 	if (pte->dma_addr)
pte               330 drivers/staging/gasket/gasket_page_table.c 		dma_unmap_page(pg_tbl->device, pte->dma_addr, PAGE_SIZE,
pte               333 drivers/staging/gasket/gasket_page_table.c 	vfree(pte->sublevel);
pte               335 drivers/staging/gasket/gasket_page_table.c 	if (pte->page)
pte               336 drivers/staging/gasket/gasket_page_table.c 		free_page((ulong)page_address(pte->page));
pte               338 drivers/staging/gasket/gasket_page_table.c 	memset(pte, 0, sizeof(struct gasket_page_table_entry));
pte               348 drivers/staging/gasket/gasket_page_table.c 	struct gasket_page_table_entry *pte;
pte               354 drivers/staging/gasket/gasket_page_table.c 	for (pte = pg_tbl->entries + pg_tbl->num_simple_entries,
pte               356 drivers/staging/gasket/gasket_page_table.c 	     pte < pg_tbl->entries + pg_tbl->config.total_entries;
pte               357 drivers/staging/gasket/gasket_page_table.c 	     pte++, slot++) {
pte               358 drivers/staging/gasket/gasket_page_table.c 		if (pte->status == PTE_INUSE) {
pte               359 drivers/staging/gasket/gasket_page_table.c 			if (gasket_is_pte_range_free(pte->sublevel,
pte               361 drivers/staging/gasket/gasket_page_table.c 				gasket_free_extended_subtable(pg_tbl, pte,
pte               646 drivers/staging/gasket/gasket_page_table.c 	struct gasket_page_table_entry *pte;
pte               651 drivers/staging/gasket/gasket_page_table.c 	pte = pg_tbl->entries + pg_tbl->num_simple_entries +
pte               658 drivers/staging/gasket/gasket_page_table.c 		if (pte->status == PTE_INUSE) {
pte               659 drivers/staging/gasket/gasket_page_table.c 			slot_base = (u64 __iomem *)(page_address(pte->page) +
pte               660 drivers/staging/gasket/gasket_page_table.c 						    pte->offset);
pte               662 drivers/staging/gasket/gasket_page_table.c 						 pte->sublevel + slot_idx,
pte               668 drivers/staging/gasket/gasket_page_table.c 		pte++;
pte               848 drivers/staging/gasket/gasket_page_table.c 					  struct gasket_page_table_entry *pte,
pte               863 drivers/staging/gasket/gasket_page_table.c 	pte->page = virt_to_page((void *)page_addr);
pte               864 drivers/staging/gasket/gasket_page_table.c 	pte->offset = 0;
pte               868 drivers/staging/gasket/gasket_page_table.c 	pte->sublevel = vzalloc(subtable_bytes);
pte               869 drivers/staging/gasket/gasket_page_table.c 	if (!pte->sublevel) {
pte               871 drivers/staging/gasket/gasket_page_table.c 		memset(pte, 0, sizeof(struct gasket_page_table_entry));
pte               876 drivers/staging/gasket/gasket_page_table.c 	pte->dma_addr = dma_map_page(pg_tbl->device, pte->page, 0, PAGE_SIZE,
pte               878 drivers/staging/gasket/gasket_page_table.c 	if (dma_mapping_error(pg_tbl->device, pte->dma_addr)) {
pte               880 drivers/staging/gasket/gasket_page_table.c 		vfree(pte->sublevel);
pte               881 drivers/staging/gasket/gasket_page_table.c 		memset(pte, 0, sizeof(struct gasket_page_table_entry));
pte               886 drivers/staging/gasket/gasket_page_table.c 	dma_addr = (pte->dma_addr + pte->offset) | GASKET_VALID_SLOT_FLAG;
pte               889 drivers/staging/gasket/gasket_page_table.c 	pte->status = PTE_INUSE;
pte               913 drivers/staging/gasket/gasket_page_table.c 	struct gasket_page_table_entry *pte;
pte               918 drivers/staging/gasket/gasket_page_table.c 	pte = pg_tbl->entries + pg_tbl->num_simple_entries +
pte               927 drivers/staging/gasket/gasket_page_table.c 		if (pte->status == PTE_FREE) {
pte               928 drivers/staging/gasket/gasket_page_table.c 			ret = gasket_alloc_extended_subtable(pg_tbl, pte, slot);
pte               935 drivers/staging/gasket/gasket_page_table.c 			if (!gasket_is_pte_range_free(pte->sublevel +
pte               942 drivers/staging/gasket/gasket_page_table.c 		pte++;
pte               960 drivers/staging/gasket/gasket_page_table.c 	struct gasket_page_table_entry *pte;
pte               978 drivers/staging/gasket/gasket_page_table.c 	pte = pg_tbl->entries + pg_tbl->num_simple_entries +
pte               985 drivers/staging/gasket/gasket_page_table.c 			(u64 __iomem *)(page_address(pte->page) + pte->offset);
pte               986 drivers/staging/gasket/gasket_page_table.c 		ret = gasket_perform_mapping(pg_tbl, pte->sublevel + slot_idx,
pte               997 drivers/staging/gasket/gasket_page_table.c 		pte++;
pte              1093 drivers/staging/gasket/gasket_page_table.c 	struct gasket_page_table_entry *pte;
pte              1101 drivers/staging/gasket/gasket_page_table.c 		pte = pg_tbl->entries + page_num;
pte              1102 drivers/staging/gasket/gasket_page_table.c 		if (pte->status != PTE_INUSE)
pte              1110 drivers/staging/gasket/gasket_page_table.c 		pte = pg_tbl->entries + pg_tbl->num_simple_entries + page_num;
pte              1111 drivers/staging/gasket/gasket_page_table.c 		if (pte->status != PTE_INUSE)
pte              1116 drivers/staging/gasket/gasket_page_table.c 		pte = pte->sublevel + page_num;
pte              1117 drivers/staging/gasket/gasket_page_table.c 		if (pte->status != PTE_INUSE)
pte              1121 drivers/staging/gasket/gasket_page_table.c 	*ppage = pte->page;
pte              1122 drivers/staging/gasket/gasket_page_table.c 	*poffset = pte->offset;
pte                29 drivers/staging/media/ipu3/ipu3-mmu.c #define IPU3_PTE2ADDR(pte)	((phys_addr_t)(pte) << IPU3_PAGE_SHIFT)
pte               124 drivers/staging/media/ipu3/ipu3-mmu.c 	int pte;
pte               130 drivers/staging/media/ipu3/ipu3-mmu.c 	for (pte = 0; pte < IPU3_PT_PTES; pte++)
pte               131 drivers/staging/media/ipu3/ipu3-mmu.c 		pt[pte] = pteval;
pte              1247 drivers/target/sbp/sbp_target.c 	struct sbp_page_table_entry *pte;
pte              1276 drivers/target/sbp/sbp_target.c 		pte = req->pg_tbl;
pte              1282 drivers/target/sbp/sbp_target.c 		pte = NULL;
pte              1294 drivers/target/sbp/sbp_target.c 			offset = (u64)be16_to_cpu(pte->segment_base_hi) << 32 |
pte              1295 drivers/target/sbp/sbp_target.c 				be32_to_cpu(pte->segment_base_lo);
pte              1296 drivers/target/sbp/sbp_target.c 			length = be16_to_cpu(pte->segment_length);
pte              1298 drivers/target/sbp/sbp_target.c 			pte++;
pte               265 drivers/xen/gntdev.c static int find_grant_ptes(pte_t *pte, unsigned long addr, void *data)
pte               273 drivers/xen/gntdev.c 	pte_maddr = arbitrary_virt_to_machine(pte).maddr;
pte               292 drivers/xen/gntdev.c static int set_grant_ptes_as_special(pte_t *pte, unsigned long addr, void *data)
pte               294 drivers/xen/gntdev.c 	set_pte_at(current->mm, addr, pte, pte_mkspecial(*pte));
pte               943 drivers/xen/privcmd.c static int is_mapped_fn(pte_t *pte, unsigned long addr, void *data)
pte               945 drivers/xen/privcmd.c 	return pte_none(*pte) ? 0 : -EBUSY;
pte               100 drivers/xen/xlate_mmu.c 	pte_t pte = pte_mkspecial(pfn_pte(page_to_pfn(page), info->prot));
pte               138 drivers/xen/xlate_mmu.c 		set_pte_at(info->vma->vm_mm, addr, ptep, pte);
pte               277 drivers/xen/xlate_mmu.c 	pte_t pte = pte_mkspecial(pfn_pte(page_to_pfn(page), r->prot));
pte               279 drivers/xen/xlate_mmu.c 	set_pte_at(r->mm, addr, ptep, pte);
pte               780 fs/dax.c       	pte_t pte, *ptep = NULL;
pte               836 fs/dax.c       			pte = ptep_clear_flush(vma, address, ptep);
pte               837 fs/dax.c       			pte = pte_wrprotect(pte);
pte               838 fs/dax.c       			pte = pte_mkclean(pte);
pte               839 fs/dax.c       			set_pte_at(vma->vm_mm, address, ptep, pte);
pte               521 fs/proc/task_mmu.c static void smaps_pte_entry(pte_t *pte, unsigned long addr,
pte               529 fs/proc/task_mmu.c 	if (pte_present(*pte)) {
pte               530 fs/proc/task_mmu.c 		page = vm_normal_page(vma, addr, *pte);
pte               531 fs/proc/task_mmu.c 	} else if (is_swap_pte(*pte)) {
pte               532 fs/proc/task_mmu.c 		swp_entry_t swpent = pte_to_swp_entry(*pte);
pte               552 fs/proc/task_mmu.c 							&& pte_none(*pte))) {
pte               569 fs/proc/task_mmu.c 	smaps_account(mss, page, false, pte_young(*pte), pte_dirty(*pte), locked);
pte               606 fs/proc/task_mmu.c 	pte_t *pte;
pte               624 fs/proc/task_mmu.c 	pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl);
pte               625 fs/proc/task_mmu.c 	for (; addr != end; pte++, addr += PAGE_SIZE)
pte               626 fs/proc/task_mmu.c 		smaps_pte_entry(pte, addr, walk);
pte               627 fs/proc/task_mmu.c 	pte_unmap_unlock(pte - 1, ptl);
pte               707 fs/proc/task_mmu.c static int smaps_hugetlb_range(pte_t *pte, unsigned long hmask,
pte               715 fs/proc/task_mmu.c 	if (pte_present(*pte)) {
pte               716 fs/proc/task_mmu.c 		page = vm_normal_page(vma, addr, *pte);
pte               717 fs/proc/task_mmu.c 	} else if (is_swap_pte(*pte)) {
pte               718 fs/proc/task_mmu.c 		swp_entry_t swpent = pte_to_swp_entry(*pte);
pte               989 fs/proc/task_mmu.c 		unsigned long addr, pte_t *pte)
pte               997 fs/proc/task_mmu.c 	pte_t ptent = *pte;
pte              1002 fs/proc/task_mmu.c 		old_pte = ptep_modify_prot_start(vma, addr, pte);
pte              1005 fs/proc/task_mmu.c 		ptep_modify_prot_commit(vma, addr, pte, old_pte, ptent);
pte              1008 fs/proc/task_mmu.c 		set_pte_at(vma->vm_mm, addr, pte, ptent);
pte              1013 fs/proc/task_mmu.c 		unsigned long addr, pte_t *pte)
pte              1053 fs/proc/task_mmu.c 	pte_t *pte, ptent;
pte              1081 fs/proc/task_mmu.c 	pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl);
pte              1082 fs/proc/task_mmu.c 	for (; addr != end; pte++, addr += PAGE_SIZE) {
pte              1083 fs/proc/task_mmu.c 		ptent = *pte;
pte              1086 fs/proc/task_mmu.c 			clear_soft_dirty(vma, addr, pte);
pte              1098 fs/proc/task_mmu.c 		ptep_test_and_clear_young(vma, addr, pte);
pte              1102 fs/proc/task_mmu.c 	pte_unmap_unlock(pte - 1, ptl);
pte              1325 fs/proc/task_mmu.c 		struct vm_area_struct *vma, unsigned long addr, pte_t pte)
pte              1330 fs/proc/task_mmu.c 	if (pte_present(pte)) {
pte              1332 fs/proc/task_mmu.c 			frame = pte_pfn(pte);
pte              1334 fs/proc/task_mmu.c 		page = vm_normal_page(vma, addr, pte);
pte              1335 fs/proc/task_mmu.c 		if (pte_soft_dirty(pte))
pte              1337 fs/proc/task_mmu.c 	} else if (is_swap_pte(pte)) {
pte              1339 fs/proc/task_mmu.c 		if (pte_swp_soft_dirty(pte))
pte              1341 fs/proc/task_mmu.c 		entry = pte_to_swp_entry(pte);
pte              1369 fs/proc/task_mmu.c 	pte_t *pte, *orig_pte;
pte              1439 fs/proc/task_mmu.c 	orig_pte = pte = pte_offset_map_lock(walk->mm, pmdp, addr, &ptl);
pte              1440 fs/proc/task_mmu.c 	for (; addr < end; pte++, addr += PAGE_SIZE) {
pte              1443 fs/proc/task_mmu.c 		pme = pte_to_pagemap_entry(pm, vma, addr, *pte);
pte              1465 fs/proc/task_mmu.c 	pte_t pte;
pte              1470 fs/proc/task_mmu.c 	pte = huge_ptep_get(ptep);
pte              1471 fs/proc/task_mmu.c 	if (pte_present(pte)) {
pte              1472 fs/proc/task_mmu.c 		struct page *page = pte_page(pte);
pte              1482 fs/proc/task_mmu.c 			frame = pte_pfn(pte) +
pte              1694 fs/proc/task_mmu.c static struct page *can_gather_numa_stats(pte_t pte, struct vm_area_struct *vma,
pte              1700 fs/proc/task_mmu.c 	if (!pte_present(pte))
pte              1703 fs/proc/task_mmu.c 	page = vm_normal_page(vma, addr, pte);
pte              1750 fs/proc/task_mmu.c 	pte_t *pte;
pte              1768 fs/proc/task_mmu.c 	orig_pte = pte = pte_offset_map_lock(walk->mm, pmd, addr, &ptl);
pte              1770 fs/proc/task_mmu.c 		struct page *page = can_gather_numa_stats(*pte, vma, addr);
pte              1773 fs/proc/task_mmu.c 		gather_stats(page, md, pte_dirty(*pte), 1);
pte              1775 fs/proc/task_mmu.c 	} while (pte++, addr += PAGE_SIZE, addr != end);
pte              1781 fs/proc/task_mmu.c static int gather_hugetlb_stats(pte_t *pte, unsigned long hmask,
pte              1784 fs/proc/task_mmu.c 	pte_t huge_pte = huge_ptep_get(pte);
pte              1801 fs/proc/task_mmu.c static int gather_hugetlb_stats(pte_t *pte, unsigned long hmask,
pte               234 fs/userfaultfd.c 	pte_t *ptep, pte;
pte               245 fs/userfaultfd.c 	pte = huge_ptep_get(ptep);
pte               251 fs/userfaultfd.c 	if (huge_pte_none(pte))
pte               253 fs/userfaultfd.c 	if (!huge_pte_write(pte) && (reason & VM_UFFD_WP))
pte               286 fs/userfaultfd.c 	pte_t *pte;
pte               324 fs/userfaultfd.c 	pte = pte_offset_map(pmd, address);
pte               329 fs/userfaultfd.c 	if (pte_none(*pte))
pte               331 fs/userfaultfd.c 	pte_unmap(pte);
pte                10 include/asm-generic/hugetlb.h static inline unsigned long huge_pte_write(pte_t pte)
pte                12 include/asm-generic/hugetlb.h 	return pte_write(pte);
pte                15 include/asm-generic/hugetlb.h static inline unsigned long huge_pte_dirty(pte_t pte)
pte                17 include/asm-generic/hugetlb.h 	return pte_dirty(pte);
pte                20 include/asm-generic/hugetlb.h static inline pte_t huge_pte_mkwrite(pte_t pte)
pte                22 include/asm-generic/hugetlb.h 	return pte_mkwrite(pte);
pte                25 include/asm-generic/hugetlb.h static inline pte_t huge_pte_mkdirty(pte_t pte)
pte                27 include/asm-generic/hugetlb.h 	return pte_mkdirty(pte);
pte                30 include/asm-generic/hugetlb.h static inline pte_t huge_pte_modify(pte_t pte, pgprot_t newprot)
pte                32 include/asm-generic/hugetlb.h 	return pte_modify(pte, newprot);
pte                54 include/asm-generic/hugetlb.h 		pte_t *ptep, pte_t pte)
pte                56 include/asm-generic/hugetlb.h 	set_pte_at(mm, addr, ptep, pte);
pte                77 include/asm-generic/hugetlb.h static inline int huge_pte_none(pte_t pte)
pte                79 include/asm-generic/hugetlb.h 	return pte_none(pte);
pte                84 include/asm-generic/hugetlb.h static inline pte_t huge_pte_wrprotect(pte_t pte)
pte                86 include/asm-generic/hugetlb.h 	return pte_wrprotect(pte);
pte               116 include/asm-generic/hugetlb.h 		pte_t pte, int dirty)
pte               118 include/asm-generic/hugetlb.h 	return ptep_set_access_flags(vma, addr, ptep, pte, dirty);
pte                38 include/asm-generic/page.h 	unsigned long pte;
pte                51 include/asm-generic/page.h #define pte_val(x)	((x).pte)
pte                42 include/asm-generic/pgalloc.h static inline void pte_free_kernel(struct mm_struct *mm, pte_t *pte)
pte                44 include/asm-generic/pgalloc.h 	free_page((unsigned long)pte);
pte                61 include/asm-generic/pgalloc.h 	struct page *pte;
pte                63 include/asm-generic/pgalloc.h 	pte = alloc_page(gfp);
pte                64 include/asm-generic/pgalloc.h 	if (!pte)
pte                66 include/asm-generic/pgalloc.h 	if (!pgtable_pte_page_ctor(pte)) {
pte                67 include/asm-generic/pgalloc.h 		__free_page(pte);
pte                71 include/asm-generic/pgalloc.h 	return pte;
pte                36 include/asm-generic/pgtable-nopmd.h #define pud_populate(mm, pmd, pte)		do { } while (0)
pte                66 include/asm-generic/pgtable.h 	pte_t pte = *ptep;
pte                68 include/asm-generic/pgtable.h 	if (!pte_young(pte))
pte                71 include/asm-generic/pgtable.h 		set_pte_at(vma->vm_mm, address, ptep, pte_mkold(pte));
pte               129 include/asm-generic/pgtable.h 	pte_t pte = *ptep;
pte               131 include/asm-generic/pgtable.h 	return pte;
pte               184 include/asm-generic/pgtable.h 	pte_t pte;
pte               185 include/asm-generic/pgtable.h 	pte = ptep_get_and_clear(mm, address, ptep);
pte               186 include/asm-generic/pgtable.h 	return pte;
pte               346 include/asm-generic/pgtable.h static inline int pte_unused(pte_t pte)
pte               353 include/asm-generic/pgtable.h #define pte_access_permitted(pte, write) \
pte               354 include/asm-generic/pgtable.h 	(pte_present(pte) && (!(write) || pte_write(pte)))
pte               411 include/asm-generic/pgtable.h #define set_pte_safe(ptep, pte) \
pte               413 include/asm-generic/pgtable.h 	WARN_ON_ONCE(pte_present(*ptep) && !pte_same(*ptep, pte)); \
pte               414 include/asm-generic/pgtable.h 	set_pte(ptep, pte); \
pte               453 include/asm-generic/pgtable.h 				     pte_t pte, pte_t oldpte)
pte               482 include/asm-generic/pgtable.h #define move_pte(pte, prot, old_addr, new_addr)	(pte)
pte               486 include/asm-generic/pgtable.h # define pte_accessible(mm, pte)	((void)(pte), 1)
pte               623 include/asm-generic/pgtable.h 					     pte_t *ptep, pte_t pte)
pte               629 include/asm-generic/pgtable.h 	set_pte_at(vma->vm_mm, addr, ptep, pte);
pte               660 include/asm-generic/pgtable.h 					   pte_t *ptep, pte_t old_pte, pte_t pte)
pte               662 include/asm-generic/pgtable.h 	__ptep_modify_prot_commit(vma, addr, ptep, pte);
pte               733 include/asm-generic/pgtable.h static inline int pte_soft_dirty(pte_t pte)
pte               743 include/asm-generic/pgtable.h static inline pte_t pte_mksoft_dirty(pte_t pte)
pte               745 include/asm-generic/pgtable.h 	return pte;
pte               753 include/asm-generic/pgtable.h static inline pte_t pte_clear_soft_dirty(pte_t pte)
pte               755 include/asm-generic/pgtable.h 	return pte;
pte               763 include/asm-generic/pgtable.h static inline pte_t pte_swp_mksoft_dirty(pte_t pte)
pte               765 include/asm-generic/pgtable.h 	return pte;
pte               768 include/asm-generic/pgtable.h static inline int pte_swp_soft_dirty(pte_t pte)
pte               773 include/asm-generic/pgtable.h static inline pte_t pte_swp_clear_soft_dirty(pte_t pte)
pte               775 include/asm-generic/pgtable.h 	return pte;
pte              1026 include/asm-generic/pgtable.h static inline int pte_protnone(pte_t pte)
pte               142 include/linux/hugetlb.h bool is_hugetlb_entry_migration(pte_t pte);
pte               537 include/linux/hugetlb.h 					   struct mm_struct *mm, pte_t *pte)
pte               540 include/linux/hugetlb.h 		return pmd_lockptr(mm, (pmd_t *) pte);
pte               568 include/linux/hugetlb.h 					pte_t *ptep, pte_t pte, unsigned long sz)
pte               570 include/linux/hugetlb.h 	set_huge_pte_at(mm, addr, ptep, pte);
pte               587 include/linux/hugetlb.h 						pte_t old_pte, pte_t pte)
pte               589 include/linux/hugetlb.h 	set_huge_pte_at(vma->vm_mm, addr, ptep, pte);
pte               723 include/linux/hugetlb.h 					   struct mm_struct *mm, pte_t *pte)
pte               737 include/linux/hugetlb.h 					pte_t *ptep, pte_t pte, unsigned long sz)
pte               743 include/linux/hugetlb.h 					struct mm_struct *mm, pte_t *pte)
pte               747 include/linux/hugetlb.h 	ptl = huge_pte_lockptr(h, mm, pte);
pte               606 include/linux/intel-iommu.h static inline void dma_clear_pte(struct dma_pte *pte)
pte               608 include/linux/intel-iommu.h 	pte->val = 0;
pte               611 include/linux/intel-iommu.h static inline u64 dma_pte_addr(struct dma_pte *pte)
pte               614 include/linux/intel-iommu.h 	return pte->val & VTD_PAGE_MASK;
pte               617 include/linux/intel-iommu.h 	return  __cmpxchg64(&pte->val, 0ULL, 0ULL) & VTD_PAGE_MASK;
pte               621 include/linux/intel-iommu.h static inline bool dma_pte_present(struct dma_pte *pte)
pte               623 include/linux/intel-iommu.h 	return (pte->val & 3) != 0;
pte               626 include/linux/intel-iommu.h static inline bool dma_pte_superpage(struct dma_pte *pte)
pte               628 include/linux/intel-iommu.h 	return (pte->val & DMA_PTE_LARGE_PAGE);
pte               631 include/linux/intel-iommu.h static inline int first_pte_in_page(struct dma_pte *pte)
pte               633 include/linux/intel-iommu.h 	return !((unsigned long)pte & ~VTD_PAGE_MASK);
pte               438 include/linux/mm.h 	pte_t *pte;			/* Pointer to pte entry matching
pte               842 include/linux/mm.h static inline pte_t maybe_mkwrite(pte_t pte, struct vm_area_struct *vma)
pte               845 include/linux/mm.h 		pte = pte_mkwrite(pte);
pte               846 include/linux/mm.h 	return pte;
pte              1452 include/linux/mm.h 			     pte_t pte);
pte              1742 include/linux/mm.h static inline int pte_devmap(pte_t pte)
pte              1984 include/linux/mm.h #define pte_unmap_unlock(pte, ptl)	do {		\
pte              1986 include/linux/mm.h 	pte_unmap(pte);					\
pte              2643 include/linux/mm.h typedef int (*pte_fn_t)(pte_t *pte, unsigned long addr, void *data);
pte               141 include/linux/mmu_notifier.h 			   pte_t pte);
pte               289 include/linux/mmu_notifier.h 				      unsigned long address, pte_t pte);
pte               337 include/linux/mmu_notifier.h 					   unsigned long address, pte_t pte)
pte               340 include/linux/mmu_notifier.h 		__mmu_notifier_change_pte(mm, address, pte);
pte               576 include/linux/mmu_notifier.h 					   unsigned long address, pte_t pte)
pte                33 include/linux/pagewalk.h 	int (*pte_entry)(pte_t *pte, unsigned long addr,
pte                37 include/linux/pagewalk.h 	int (*hugetlb_entry)(pte_t *pte, unsigned long hmask,
pte               112 include/linux/pfn_t.h pte_t pte_mkdevmap(pte_t pte);
pte               210 include/linux/rmap.h 	pte_t *pte;
pte               217 include/linux/rmap.h 	if (pvmw->pte)
pte               218 include/linux/rmap.h 		pte_unmap(pvmw->pte);
pte                56 include/linux/swapops.h static inline int is_swap_pte(pte_t pte)
pte                58 include/linux/swapops.h 	return !pte_none(pte) && !pte_present(pte);
pte                65 include/linux/swapops.h static inline swp_entry_t pte_to_swp_entry(pte_t pte)
pte                69 include/linux/swapops.h 	if (pte_swp_soft_dirty(pte))
pte                70 include/linux/swapops.h 		pte = pte_swp_clear_soft_dirty(pte);
pte                71 include/linux/swapops.h 	arch_entry = __pte_to_swp_entry(pte);
pte               209 include/linux/swapops.h 		struct mm_struct *mm, pte_t *pte);
pte               234 include/linux/swapops.h 		struct mm_struct *mm, pte_t *pte) { }
pte                13 include/trace/events/thp.h 	    TP_PROTO(unsigned long addr, unsigned long pte),
pte                14 include/trace/events/thp.h 	    TP_ARGS(addr, pte),
pte                17 include/trace/events/thp.h 		    __field(unsigned long, pte)
pte                22 include/trace/events/thp.h 		    __entry->pte = pte;
pte                26 include/trace/events/thp.h 		      __entry->addr, __entry->pte)
pte                49 include/trace/events/thp.h 	    TP_PROTO(unsigned long addr, unsigned long pte, unsigned long clr, unsigned long set),
pte                50 include/trace/events/thp.h 	    TP_ARGS(addr, pte, clr, set),
pte                53 include/trace/events/thp.h 		    __field(unsigned long, pte)
pte                60 include/trace/events/thp.h 		    __entry->pte = pte;
pte                66 include/trace/events/thp.h 	    TP_printk("hugepage update at addr 0x%lx and pte = 0x%lx clr = 0x%lx, set = 0x%lx", __entry->addr, __entry->pte, __entry->clr, __entry->set)
pte                70 include/trace/events/thp.h 	    TP_PROTO(unsigned long addr, unsigned long pte),
pte                71 include/trace/events/thp.h 	    TP_ARGS(addr, pte),
pte                74 include/trace/events/thp.h 		    __field(unsigned long, pte)
pte                79 include/trace/events/thp.h 		    __entry->pte = pte;
pte                83 include/trace/events/thp.h 		      __entry->addr, __entry->pte)
pte               142 include/trace/events/xen.h 			   __entry->pteval = pteval.pte),
pte               169 include/trace/events/xen.h 			   __entry->pteval = pteval.pte),
pte               291 include/trace/events/xen.h 			   __entry->pteval = pteval.pte),
pte               203 kernel/events/uprobes.c 	flush_cache_page(vma, addr, pte_pfn(*pvmw.pte));
pte               204 kernel/events/uprobes.c 	ptep_clear_flush_notify(vma, addr, pvmw.pte);
pte               206 kernel/events/uprobes.c 		set_pte_at_notify(mm, addr, pvmw.pte,
pte                66 lib/ioremap.c  	pte_t *pte;
pte                70 lib/ioremap.c  	pte = pte_alloc_kernel(pmd, addr);
pte                71 lib/ioremap.c  	if (!pte)
pte                74 lib/ioremap.c  		BUG_ON(!pte_none(*pte));
pte                75 lib/ioremap.c  		set_pte_at(&init_mm, addr, pte, pfn_pte(pfn, prot));
pte                77 lib/ioremap.c  	} while (pte++, addr += PAGE_SIZE, addr != end);
pte              2651 mm/filemap.c   		if (vmf->pte)
pte              2652 mm/filemap.c   			vmf->pte += xas.xa_index - last_pgoff;
pte               140 mm/gup.c       		pte_t *pte, unsigned int flags)
pte               147 mm/gup.c       		pte_t entry = *pte;
pte               153 mm/gup.c       		if (!pte_same(*pte, entry)) {
pte               154 mm/gup.c       			set_pte_at(vma->vm_mm, address, pte, entry);
pte               155 mm/gup.c       			update_mmu_cache(vma, address, pte);
pte               167 mm/gup.c       static inline bool can_follow_write_pte(pte_t pte, unsigned int flags)
pte               169 mm/gup.c       	return pte_write(pte) || ((flags & FOLL_COW) && pte_dirty(pte));
pte               189 mm/gup.c       	pte_t *ptep, pte;
pte               196 mm/gup.c       	pte = *ptep;
pte               197 mm/gup.c       	if (!pte_present(pte)) {
pte               206 mm/gup.c       		if (pte_none(pte))
pte               208 mm/gup.c       		entry = pte_to_swp_entry(pte);
pte               215 mm/gup.c       	if ((flags & FOLL_NUMA) && pte_protnone(pte))
pte               217 mm/gup.c       	if ((flags & FOLL_WRITE) && !can_follow_write_pte(pte, flags)) {
pte               222 mm/gup.c       	page = vm_normal_page(vma, address, pte);
pte               223 mm/gup.c       	if (!page && pte_devmap(pte) && (flags & FOLL_GET)) {
pte               228 mm/gup.c       		*pgmap = get_dev_pagemap(pte_pfn(pte), *pgmap);
pte               230 mm/gup.c       			page = pte_page(pte);
pte               240 mm/gup.c       		if (is_zero_pfn(pte_pfn(pte))) {
pte               241 mm/gup.c       			page = pte_page(pte);
pte               272 mm/gup.c       		    !pte_dirty(pte) && !PageDirty(page))
pte               312 mm/gup.c       	if (!pte_none(pte))
pte               577 mm/gup.c       	pte_t *pte;
pte               599 mm/gup.c       	pte = pte_offset_map(pmd, address);
pte               600 mm/gup.c       	if (pte_none(*pte))
pte               605 mm/gup.c       	*page = vm_normal_page(*vma, address, *pte);
pte               607 mm/gup.c       		if ((gup_flags & FOLL_DUMP) || !is_zero_pfn(pte_pfn(*pte)))
pte               609 mm/gup.c       		*page = pte_page(*pte);
pte               618 mm/gup.c       	pte_unmap(pte);
pte              1779 mm/gup.c       	pte_t pte;
pte              1782 mm/gup.c       		pte.pte_low = ptep->pte_low;
pte              1784 mm/gup.c       		pte.pte_high = ptep->pte_high;
pte              1786 mm/gup.c       	} while (unlikely(pte.pte_low != ptep->pte_low));
pte              1788 mm/gup.c       	return pte;
pte              1835 mm/gup.c       		pte_t pte = gup_get_pte(ptep);
pte              1842 mm/gup.c       		if (pte_protnone(pte))
pte              1845 mm/gup.c       		if (!pte_access_permitted(pte, flags & FOLL_WRITE))
pte              1848 mm/gup.c       		if (pte_devmap(pte)) {
pte              1852 mm/gup.c       			pgmap = get_dev_pagemap(pte_pfn(pte), pgmap);
pte              1857 mm/gup.c       		} else if (pte_special(pte))
pte              1860 mm/gup.c       		VM_BUG_ON(!pfn_valid(pte_pfn(pte)));
pte              1861 mm/gup.c       		page = pte_page(pte);
pte              1867 mm/gup.c       		if (unlikely(pte_val(pte) != pte_val(*ptep))) {
pte              1996 mm/gup.c       	pte_t pte;
pte              2003 mm/gup.c       	pte = READ_ONCE(*ptep);
pte              2005 mm/gup.c       	if (!pte_access_permitted(pte, flags & FOLL_WRITE))
pte              2009 mm/gup.c       	VM_BUG_ON(!pfn_valid(pte_pfn(pte)));
pte              2012 mm/gup.c       	head = pte_page(pte);
pte              2029 mm/gup.c       	if (unlikely(pte_val(pte) != pte_val(*ptep))) {
pte               445 mm/hmm.c       static inline uint64_t pte_to_hmm_pfn_flags(struct hmm_range *range, pte_t pte)
pte               447 mm/hmm.c       	if (pte_none(pte) || !pte_present(pte) || pte_protnone(pte))
pte               449 mm/hmm.c       	return pte_write(pte) ? range->flags[HMM_PFN_VALID] |
pte               462 mm/hmm.c       	pte_t pte = *ptep;
pte               468 mm/hmm.c       	if (pte_none(pte)) {
pte               476 mm/hmm.c       	if (!pte_present(pte)) {
pte               477 mm/hmm.c       		swp_entry_t entry = pte_to_swp_entry(pte);
pte               480 mm/hmm.c       			cpu_flags = pte_to_hmm_pfn_flags(range, pte);
pte               521 mm/hmm.c       		cpu_flags = pte_to_hmm_pfn_flags(range, pte);
pte               529 mm/hmm.c       	if (pte_devmap(pte)) {
pte               530 mm/hmm.c       		hmm_vma_walk->pgmap = get_dev_pagemap(pte_pfn(pte),
pte               534 mm/hmm.c       	} else if (IS_ENABLED(CONFIG_ARCH_HAS_PTE_SPECIAL) && pte_special(pte)) {
pte               539 mm/hmm.c       	*pfn = hmm_device_entry_from_pfn(range, pte_pfn(pte)) | cpu_flags;
pte               726 mm/hmm.c       static int hmm_vma_walk_hugetlb_entry(pte_t *pte, unsigned long hmask,
pte               740 mm/hmm.c       	ptl = huge_pte_lock(hstate_vma(vma), walk->mm, pte);
pte               741 mm/hmm.c       	entry = huge_ptep_get(pte);
pte              1278 mm/huge_memory.c 		vmf->pte = pte_offset_map(&_pmd, haddr);
pte              1279 mm/huge_memory.c 		VM_BUG_ON(!pte_none(*vmf->pte));
pte              1280 mm/huge_memory.c 		set_pte_at(vma->vm_mm, haddr, vmf->pte, entry);
pte              1281 mm/huge_memory.c 		pte_unmap(vmf->pte);
pte              2127 mm/huge_memory.c 		pte_t *pte, entry;
pte              2130 mm/huge_memory.c 		pte = pte_offset_map(&_pmd, haddr);
pte              2131 mm/huge_memory.c 		VM_BUG_ON(!pte_none(*pte));
pte              2132 mm/huge_memory.c 		set_pte_at(mm, haddr, pte, entry);
pte              2133 mm/huge_memory.c 		pte_unmap(pte);
pte              2240 mm/huge_memory.c 		pte_t entry, *pte;
pte              2262 mm/huge_memory.c 		pte = pte_offset_map(&_pmd, addr);
pte              2263 mm/huge_memory.c 		BUG_ON(!pte_none(*pte));
pte              2264 mm/huge_memory.c 		set_pte_at(mm, addr, pte, entry);
pte              2266 mm/huge_memory.c 		pte_unmap(pte);
pte              3030 mm/huge_memory.c 	if (!(pvmw->pmd && !pvmw->pte))
pte              3055 mm/huge_memory.c 	if (!(pvmw->pmd && !pvmw->pte))
pte              3393 mm/hugetlb.c   bool is_hugetlb_entry_migration(pte_t pte)
pte              3397 mm/hugetlb.c   	if (huge_pte_none(pte) || pte_present(pte))
pte              3399 mm/hugetlb.c   	swp = pte_to_swp_entry(pte);
pte              3406 mm/hugetlb.c   static int is_hugetlb_entry_hwpoisoned(pte_t pte)
pte              3410 mm/hugetlb.c   	if (huge_pte_none(pte) || pte_present(pte))
pte              3412 mm/hugetlb.c   	swp = pte_to_swp_entry(pte);
pte              3526 mm/hugetlb.c   	pte_t pte;
pte              3567 mm/hugetlb.c   		pte = huge_ptep_get(ptep);
pte              3568 mm/hugetlb.c   		if (huge_pte_none(pte)) {
pte              3577 mm/hugetlb.c   		if (unlikely(!pte_present(pte))) {
pte              3583 mm/hugetlb.c   		page = pte_page(pte);
pte              3602 mm/hugetlb.c   		pte = huge_ptep_get_and_clear(mm, address, ptep);
pte              3604 mm/hugetlb.c   		if (huge_pte_dirty(pte))
pte              3731 mm/hugetlb.c   	pte_t pte;
pte              3739 mm/hugetlb.c   	pte = huge_ptep_get(ptep);
pte              3740 mm/hugetlb.c   	old_page = pte_page(pte);
pte              3783 mm/hugetlb.c   			BUG_ON(huge_pte_none(pte));
pte              3785 mm/hugetlb.c   			BUG_ON(huge_pte_none(pte));
pte              3789 mm/hugetlb.c   				   pte_same(huge_ptep_get(ptep), pte)))
pte              3825 mm/hugetlb.c   	if (likely(ptep && pte_same(huge_ptep_get(ptep), pte))) {
pte              4399 mm/hugetlb.c   		pte_t *pte;
pte              4420 mm/hugetlb.c   		pte = huge_pte_offset(mm, vaddr & huge_page_mask(h),
pte              4422 mm/hugetlb.c   		if (pte)
pte              4423 mm/hugetlb.c   			ptl = huge_pte_lock(h, mm, pte);
pte              4424 mm/hugetlb.c   		absent = !pte || huge_pte_none(huge_ptep_get(pte));
pte              4435 mm/hugetlb.c   			if (pte)
pte              4451 mm/hugetlb.c   		if (absent || is_swap_pte(huge_ptep_get(pte)) ||
pte              4453 mm/hugetlb.c   		      !huge_pte_write(huge_ptep_get(pte)))) {
pte              4457 mm/hugetlb.c   			if (pte)
pte              4497 mm/hugetlb.c   		page = pte_page(huge_ptep_get(pte));
pte              4559 mm/hugetlb.c   	pte_t pte;
pte              4591 mm/hugetlb.c   		pte = huge_ptep_get(ptep);
pte              4592 mm/hugetlb.c   		if (unlikely(is_hugetlb_entry_hwpoisoned(pte))) {
pte              4596 mm/hugetlb.c   		if (unlikely(is_hugetlb_entry_migration(pte))) {
pte              4597 mm/hugetlb.c   			swp_entry_t entry = pte_to_swp_entry(pte);
pte              4611 mm/hugetlb.c   		if (!huge_pte_none(pte)) {
pte              4615 mm/hugetlb.c   			pte = pte_mkhuge(huge_pte_modify(old_pte, newprot));
pte              4616 mm/hugetlb.c   			pte = arch_make_huge_pte(pte, vma, NULL, 0);
pte              4617 mm/hugetlb.c   			huge_ptep_modify_prot_commit(vma, address, ptep, old_pte, pte);
pte              4888 mm/hugetlb.c   	pte_t *pte;
pte              4923 mm/hugetlb.c   	pte = (pte_t *)pmd_alloc(mm, pud, addr);
pte              4925 mm/hugetlb.c   	return pte;
pte              4982 mm/hugetlb.c   	pte_t *pte = NULL;
pte              4991 mm/hugetlb.c   			pte = (pte_t *)pud;
pte              4995 mm/hugetlb.c   				pte = huge_pmd_share(mm, addr, pud);
pte              4997 mm/hugetlb.c   				pte = (pte_t *)pmd_alloc(mm, pud, addr);
pte              5000 mm/hugetlb.c   	BUG_ON(pte && pte_present(*pte) && !pte_huge(*pte));
pte              5002 mm/hugetlb.c   	return pte;
pte              5075 mm/hugetlb.c   	pte_t pte;
pte              5085 mm/hugetlb.c   	pte = huge_ptep_get((pte_t *)pmd);
pte              5086 mm/hugetlb.c   	if (pte_present(pte)) {
pte              5091 mm/hugetlb.c   		if (is_hugetlb_entry_migration(pte)) {
pte               645 mm/kasan/common.c 	pte_t *pte;
pte               669 mm/kasan/common.c 	pte = pte_offset_kernel(pmd, addr);
pte               670 mm/kasan/common.c 	return !pte_none(*pte);
pte                79 mm/kasan/init.c static inline bool kasan_early_shadow_page_entry(pte_t pte)
pte                81 mm/kasan/init.c 	return pte_page(pte) == virt_to_page(lm_alias(kasan_early_shadow_page));
pte                99 mm/kasan/init.c 	pte_t *pte = pte_offset_kernel(pmd, addr);
pte               107 mm/kasan/init.c 		set_pte_at(&init_mm, addr, pte, zero_pte);
pte               109 mm/kasan/init.c 		pte = pte_offset_kernel(pmd, addr);
pte               299 mm/kasan/init.c 	pte_t *pte;
pte               303 mm/kasan/init.c 		pte = pte_start + i;
pte               304 mm/kasan/init.c 		if (!pte_none(*pte))
pte               357 mm/kasan/init.c static void kasan_remove_pte_table(pte_t *pte, unsigned long addr,
pte               362 mm/kasan/init.c 	for (; addr < end; addr = next, pte++) {
pte               367 mm/kasan/init.c 		if (!pte_present(*pte))
pte               370 mm/kasan/init.c 		if (WARN_ON(!kasan_early_shadow_page_entry(*pte)))
pte               372 mm/kasan/init.c 		pte_clear(&init_mm, addr, pte);
pte               382 mm/kasan/init.c 		pte_t *pte;
pte               395 mm/kasan/init.c 		pte = pte_offset_kernel(pmd, addr);
pte               396 mm/kasan/init.c 		kasan_remove_pte_table(pte, addr, next);
pte               523 mm/khugepaged.c static void release_pte_pages(pte_t *pte, pte_t *_pte)
pte               525 mm/khugepaged.c 	while (--_pte >= pte) {
pte               534 mm/khugepaged.c 					pte_t *pte)
pte               541 mm/khugepaged.c 	for (_pte = pte; _pte < pte+HPAGE_PMD_NR;
pte               640 mm/khugepaged.c 	release_pte_pages(pte, _pte);
pte               646 mm/khugepaged.c static void __collapse_huge_page_copy(pte_t *pte, struct page *page,
pte               652 mm/khugepaged.c 	for (_pte = pte; _pte < pte + HPAGE_PMD_NR;
pte               910 mm/khugepaged.c 	vmf.pte = pte_offset_map(pmd, address);
pte               912 mm/khugepaged.c 			vmf.pte++, vmf.address += PAGE_SIZE) {
pte               913 mm/khugepaged.c 		vmf.orig_pte = *vmf.pte;
pte               938 mm/khugepaged.c 		vmf.pte = pte_offset_map(pmd, vmf.address);
pte               940 mm/khugepaged.c 	vmf.pte--;
pte               941 mm/khugepaged.c 	pte_unmap(vmf.pte);
pte               952 mm/khugepaged.c 	pte_t *pte;
pte              1035 mm/khugepaged.c 	pte = pte_offset_map(pmd, address);
pte              1050 mm/khugepaged.c 	isolated = __collapse_huge_page_isolate(vma, address, pte);
pte              1054 mm/khugepaged.c 		pte_unmap(pte);
pte              1075 mm/khugepaged.c 	__collapse_huge_page_copy(pte, new_page, vma, address, pte_ptl);
pte              1076 mm/khugepaged.c 	pte_unmap(pte);
pte              1121 mm/khugepaged.c 	pte_t *pte, *_pte;
pte              1138 mm/khugepaged.c 	pte = pte_offset_map_lock(mm, pmd, address, &ptl);
pte              1139 mm/khugepaged.c 	for (_address = address, _pte = pte; _pte < pte+HPAGE_PMD_NR;
pte              1228 mm/khugepaged.c 	pte_unmap_unlock(pte, ptl);
pte              1295 mm/khugepaged.c 	pte_t *start_pte, *pte;
pte              1321 mm/khugepaged.c 	for (i = 0, addr = haddr, pte = start_pte;
pte              1322 mm/khugepaged.c 	     i < HPAGE_PMD_NR; i++, addr += PAGE_SIZE, pte++) {
pte              1326 mm/khugepaged.c 		if (pte_none(*pte))
pte              1330 mm/khugepaged.c 		if (!pte_present(*pte))
pte              1333 mm/khugepaged.c 		page = vm_normal_page(vma, addr, *pte);
pte              1364 mm/khugepaged.c 	for (i = 0, addr = haddr, pte = start_pte;
pte              1365 mm/khugepaged.c 	     i < HPAGE_PMD_NR; i++, addr += PAGE_SIZE, pte++) {
pte              1368 mm/khugepaged.c 		if (pte_none(*pte))
pte              1370 mm/khugepaged.c 		page = vm_normal_page(vma, addr, *pte);
pte              1057 mm/ksm.c       	if (WARN_ONCE(!pvmw.pte, "Unexpected PMD mapping?"))
pte              1060 mm/ksm.c       	if (pte_write(*pvmw.pte) || pte_dirty(*pvmw.pte) ||
pte              1061 mm/ksm.c       	    (pte_protnone(*pvmw.pte) && pte_savedwrite(*pvmw.pte)) ||
pte              1081 mm/ksm.c       		entry = ptep_clear_flush(vma, pvmw.address, pvmw.pte);
pte              1087 mm/ksm.c       			set_pte_at(mm, pvmw.address, pvmw.pte, entry);
pte              1097 mm/ksm.c       		set_pte_at_notify(mm, pvmw.address, pvmw.pte, entry);
pte              1099 mm/ksm.c       	*orig_pte = *pvmw.pte;
pte               194 mm/madvise.c   		pte_t pte;
pte               200 mm/madvise.c   		pte = *(orig_pte + ((index - start) / PAGE_SIZE));
pte               203 mm/madvise.c   		if (pte_present(pte) || pte_none(pte))
pte               205 mm/madvise.c   		entry = pte_to_swp_entry(pte);
pte               309 mm/madvise.c   	pte_t *orig_pte, *pte, ptent;
pte               388 mm/madvise.c   	orig_pte = pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl);
pte               391 mm/madvise.c   	for (; addr < end; pte++, addr += PAGE_SIZE) {
pte               392 mm/madvise.c   		ptent = *pte;
pte               425 mm/madvise.c   			pte = pte_offset_map_lock(mm, pmd, addr, &ptl);
pte               426 mm/madvise.c   			pte--;
pte               438 mm/madvise.c   			ptent = ptep_get_and_clear_full(mm, addr, pte,
pte               441 mm/madvise.c   			set_pte_at(mm, addr, pte, ptent);
pte               442 mm/madvise.c   			tlb_remove_tlb_entry(tlb, pte, addr);
pte               570 mm/madvise.c   	pte_t *orig_pte, *pte, ptent;
pte               584 mm/madvise.c   	orig_pte = pte = pte_offset_map_lock(mm, pmd, addr, &ptl);
pte               587 mm/madvise.c   	for (; addr != end; pte++, addr += PAGE_SIZE) {
pte               588 mm/madvise.c   		ptent = *pte;
pte               605 mm/madvise.c   			pte_clear_not_present_full(mm, addr, pte, tlb->fullmm);
pte               635 mm/madvise.c   			pte = pte_offset_map_lock(mm, pmd, addr, &ptl);
pte               636 mm/madvise.c   			pte--;
pte               671 mm/madvise.c   			ptent = ptep_get_and_clear_full(mm, addr, pte,
pte               676 mm/madvise.c   			set_pte_at(mm, addr, pte, ptent);
pte               677 mm/madvise.c   			tlb_remove_tlb_entry(tlb, pte, addr);
pte              5682 mm/memcontrol.c 	pte_t *pte;
pte              5700 mm/memcontrol.c 	pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl);
pte              5701 mm/memcontrol.c 	for (; addr != end; pte++, addr += PAGE_SIZE)
pte              5702 mm/memcontrol.c 		if (get_mctgt_type(vma, addr, *pte, NULL))
pte              5704 mm/memcontrol.c 	pte_unmap_unlock(pte - 1, ptl);
pte              5882 mm/memcontrol.c 	pte_t *pte;
pte              5922 mm/memcontrol.c 	pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl);
pte              5924 mm/memcontrol.c 		pte_t ptent = *(pte++);
pte              5970 mm/memcontrol.c 	pte_unmap_unlock(pte - 1, ptl);
pte               272 mm/memory-failure.c 	pte_t *pte;
pte               290 mm/memory-failure.c 	pte = pte_offset_map(pmd, address);
pte               291 mm/memory-failure.c 	if (!pte_present(*pte))
pte               293 mm/memory-failure.c 	if (pte_devmap(*pte))
pte               482 mm/memory.c    			  pte_t pte, struct page *page)
pte               518 mm/memory.c    		 (long long)pte_val(pte), (long long)pmd_val(*pmd));
pte               575 mm/memory.c    			    pte_t pte)
pte               577 mm/memory.c    	unsigned long pfn = pte_pfn(pte);
pte               580 mm/memory.c    		if (likely(!pte_special(pte)))
pte               588 mm/memory.c    		if (pte_devmap(pte))
pte               591 mm/memory.c    		print_bad_pte(vma, addr, pte, NULL);
pte               617 mm/memory.c    		print_bad_pte(vma, addr, pte, NULL);
pte               683 mm/memory.c    	pte_t pte = *src_pte;
pte               687 mm/memory.c    	if (unlikely(!pte_present(pte))) {
pte               688 mm/memory.c    		swp_entry_t entry = pte_to_swp_entry(pte);
pte               715 mm/memory.c    				pte = swp_entry_to_pte(entry);
pte               717 mm/memory.c    					pte = pte_swp_mksoft_dirty(pte);
pte               718 mm/memory.c    				set_pte_at(src_mm, addr, src_pte, pte);
pte               746 mm/memory.c    				pte = swp_entry_to_pte(entry);
pte               747 mm/memory.c    				set_pte_at(src_mm, addr, src_pte, pte);
pte               757 mm/memory.c    	if (is_cow_mapping(vm_flags) && pte_write(pte)) {
pte               759 mm/memory.c    		pte = pte_wrprotect(pte);
pte               767 mm/memory.c    		pte = pte_mkclean(pte);
pte               768 mm/memory.c    	pte = pte_mkold(pte);
pte               770 mm/memory.c    	page = vm_normal_page(vma, addr, pte);
pte               775 mm/memory.c    	} else if (pte_devmap(pte)) {
pte               776 mm/memory.c    		page = pte_page(pte);
pte               780 mm/memory.c    	set_pte_at(dst_mm, addr, dst_pte, pte);
pte              1014 mm/memory.c    	pte_t *pte;
pte              1021 mm/memory.c    	pte = start_pte;
pte              1025 mm/memory.c    		pte_t ptent = *pte;
pte              1046 mm/memory.c    			ptent = ptep_get_and_clear_full(mm, addr, pte,
pte              1048 mm/memory.c    			tlb_remove_tlb_entry(tlb, pte, addr);
pte              1088 mm/memory.c    			pte_clear_not_present_full(mm, addr, pte, tlb->fullmm);
pte              1109 mm/memory.c    		pte_clear_not_present_full(mm, addr, pte, tlb->fullmm);
pte              1110 mm/memory.c    	} while (pte++, addr += PAGE_SIZE, addr != end);
pte              1427 mm/memory.c    	pte_t *pte;
pte              1435 mm/memory.c    	pte = get_locked_pte(mm, addr, &ptl);
pte              1436 mm/memory.c    	if (!pte)
pte              1439 mm/memory.c    	if (!pte_none(*pte))
pte              1446 mm/memory.c    	set_pte_at(mm, addr, pte, mk_pte(page, prot));
pte              1450 mm/memory.c    	pte_unmap_unlock(pte, ptl);
pte              1585 mm/memory.c    	pte_t *pte, entry;
pte              1588 mm/memory.c    	pte = get_locked_pte(mm, addr, &ptl);
pte              1589 mm/memory.c    	if (!pte)
pte              1591 mm/memory.c    	if (!pte_none(*pte)) {
pte              1603 mm/memory.c    			if (pte_pfn(*pte) != pfn_t_to_pfn(pfn)) {
pte              1604 mm/memory.c    				WARN_ON_ONCE(!is_zero_pfn(pte_pfn(*pte)));
pte              1607 mm/memory.c    			entry = pte_mkyoung(*pte);
pte              1609 mm/memory.c    			if (ptep_set_access_flags(vma, addr, pte, entry, 1))
pte              1610 mm/memory.c    				update_mmu_cache(vma, addr, pte);
pte              1626 mm/memory.c    	set_pte_at(mm, addr, pte, entry);
pte              1627 mm/memory.c    	update_mmu_cache(vma, addr, pte); /* XXX: why not for insert_page? */
pte              1630 mm/memory.c    	pte_unmap_unlock(pte, ptl);
pte              1795 mm/memory.c    	pte_t *pte;
pte              1799 mm/memory.c    	pte = pte_alloc_map_lock(mm, pmd, addr, &ptl);
pte              1800 mm/memory.c    	if (!pte)
pte              1804 mm/memory.c    		BUG_ON(!pte_none(*pte));
pte              1809 mm/memory.c    		set_pte_at(mm, addr, pte, pte_mkspecial(pfn_pte(pfn, prot)));
pte              1811 mm/memory.c    	} while (pte++, addr += PAGE_SIZE, addr != end);
pte              1813 mm/memory.c    	pte_unmap_unlock(pte - 1, ptl);
pte              2008 mm/memory.c    	pte_t *pte;
pte              2012 mm/memory.c    	pte = (mm == &init_mm) ?
pte              2015 mm/memory.c    	if (!pte)
pte              2023 mm/memory.c    		err = fn(pte++, addr, data);
pte              2031 mm/memory.c    		pte_unmap_unlock(pte-1, ptl);
pte              2300 mm/memory.c    	if (ptep_set_access_flags(vma, vmf->address, vmf->pte, entry, 1))
pte              2301 mm/memory.c    		update_mmu_cache(vma, vmf->address, vmf->pte);
pte              2302 mm/memory.c    	pte_unmap_unlock(vmf->pte, vmf->ptl);
pte              2361 mm/memory.c    	vmf->pte = pte_offset_map_lock(mm, vmf->pmd, vmf->address, &vmf->ptl);
pte              2362 mm/memory.c    	if (likely(pte_same(*vmf->pte, vmf->orig_pte))) {
pte              2381 mm/memory.c    		ptep_clear_flush_notify(vma, vmf->address, vmf->pte);
pte              2390 mm/memory.c    		set_pte_at_notify(mm, vmf->address, vmf->pte, entry);
pte              2391 mm/memory.c    		update_mmu_cache(vma, vmf->address, vmf->pte);
pte              2428 mm/memory.c    	pte_unmap_unlock(vmf->pte, vmf->ptl);
pte              2475 mm/memory.c    	vmf->pte = pte_offset_map_lock(vmf->vma->vm_mm, vmf->pmd, vmf->address,
pte              2481 mm/memory.c    	if (!pte_same(*vmf->pte, vmf->orig_pte)) {
pte              2482 mm/memory.c    		pte_unmap_unlock(vmf->pte, vmf->ptl);
pte              2500 mm/memory.c    		pte_unmap_unlock(vmf->pte, vmf->ptl);
pte              2522 mm/memory.c    		pte_unmap_unlock(vmf->pte, vmf->ptl);
pte              2581 mm/memory.c    		pte_unmap_unlock(vmf->pte, vmf->ptl);
pte              2596 mm/memory.c    			pte_unmap_unlock(vmf->pte, vmf->ptl);
pte              2598 mm/memory.c    			vmf->pte = pte_offset_map_lock(vma->vm_mm, vmf->pmd,
pte              2600 mm/memory.c    			if (!pte_same(*vmf->pte, vmf->orig_pte)) {
pte              2602 mm/memory.c    				pte_unmap_unlock(vmf->pte, vmf->ptl);
pte              2643 mm/memory.c    	pte_unmap_unlock(vmf->pte, vmf->ptl);
pte              2757 mm/memory.c    	pte_t pte;
pte              2762 mm/memory.c    	if (!pte_unmap_same(vma->vm_mm, vmf->pmd, vmf->pte, vmf->orig_pte))
pte              2813 mm/memory.c    			vmf->pte = pte_offset_map_lock(vma->vm_mm, vmf->pmd,
pte              2815 mm/memory.c    			if (likely(pte_same(*vmf->pte, vmf->orig_pte)))
pte              2869 mm/memory.c    	vmf->pte = pte_offset_map_lock(vma->vm_mm, vmf->pmd, vmf->address,
pte              2871 mm/memory.c    	if (unlikely(!pte_same(*vmf->pte, vmf->orig_pte)))
pte              2891 mm/memory.c    	pte = mk_pte(page, vma->vm_page_prot);
pte              2893 mm/memory.c    		pte = maybe_mkwrite(pte_mkdirty(pte), vma);
pte              2900 mm/memory.c    		pte = pte_mksoft_dirty(pte);
pte              2901 mm/memory.c    	set_pte_at(vma->vm_mm, vmf->address, vmf->pte, pte);
pte              2902 mm/memory.c    	arch_do_swap_page(vma->vm_mm, vma, vmf->address, pte, vmf->orig_pte);
pte              2903 mm/memory.c    	vmf->orig_pte = pte;
pte              2942 mm/memory.c    	update_mmu_cache(vma, vmf->address, vmf->pte);
pte              2944 mm/memory.c    	pte_unmap_unlock(vmf->pte, vmf->ptl);
pte              2949 mm/memory.c    	pte_unmap_unlock(vmf->pte, vmf->ptl);
pte              3000 mm/memory.c    		vmf->pte = pte_offset_map_lock(vma->vm_mm, vmf->pmd,
pte              3002 mm/memory.c    		if (!pte_none(*vmf->pte))
pte              3009 mm/memory.c    			pte_unmap_unlock(vmf->pte, vmf->ptl);
pte              3037 mm/memory.c    	vmf->pte = pte_offset_map_lock(vma->vm_mm, vmf->pmd, vmf->address,
pte              3039 mm/memory.c    	if (!pte_none(*vmf->pte))
pte              3048 mm/memory.c    		pte_unmap_unlock(vmf->pte, vmf->ptl);
pte              3059 mm/memory.c    	set_pte_at(vma->vm_mm, vmf->address, vmf->pte, entry);
pte              3062 mm/memory.c    	update_mmu_cache(vma, vmf->address, vmf->pte);
pte              3064 mm/memory.c    	pte_unmap_unlock(vmf->pte, vmf->ptl);
pte              3184 mm/memory.c    	vmf->pte = pte_offset_map_lock(vma->vm_mm, vmf->pmd, vmf->address,
pte              3301 mm/memory.c    	if (!vmf->pte) {
pte              3308 mm/memory.c    	if (unlikely(!pte_none(*vmf->pte)))
pte              3325 mm/memory.c    	set_pte_at(vma->vm_mm, vmf->address, vmf->pte, entry);
pte              3328 mm/memory.c    	update_mmu_cache(vma, vmf->address, vmf->pte);
pte              3369 mm/memory.c    	if (vmf->pte)
pte              3370 mm/memory.c    		pte_unmap_unlock(vmf->pte, vmf->ptl);
pte              3475 mm/memory.c    	if (!vmf->pte)
pte              3479 mm/memory.c    	vmf->pte -= (vmf->address >> PAGE_SHIFT) - (address >> PAGE_SHIFT);
pte              3480 mm/memory.c    	if (!pte_none(*vmf->pte))
pte              3482 mm/memory.c    	pte_unmap_unlock(vmf->pte, vmf->ptl);
pte              3485 mm/memory.c    	vmf->pte = NULL;
pte              3615 mm/memory.c    			vmf->pte = pte_offset_map_lock(vmf->vma->vm_mm,
pte              3626 mm/memory.c    			if (unlikely(pte_none(*vmf->pte)))
pte              3631 mm/memory.c    			pte_unmap_unlock(vmf->pte, vmf->ptl);
pte              3671 mm/memory.c    	pte_t pte, old_pte;
pte              3682 mm/memory.c    	if (unlikely(!pte_same(*vmf->pte, vmf->orig_pte))) {
pte              3683 mm/memory.c    		pte_unmap_unlock(vmf->pte, vmf->ptl);
pte              3691 mm/memory.c    	old_pte = ptep_modify_prot_start(vma, vmf->address, vmf->pte);
pte              3692 mm/memory.c    	pte = pte_modify(old_pte, vma->vm_page_prot);
pte              3693 mm/memory.c    	pte = pte_mkyoung(pte);
pte              3695 mm/memory.c    		pte = pte_mkwrite(pte);
pte              3696 mm/memory.c    	ptep_modify_prot_commit(vma, vmf->address, vmf->pte, old_pte, pte);
pte              3697 mm/memory.c    	update_mmu_cache(vma, vmf->address, vmf->pte);
pte              3699 mm/memory.c    	page = vm_normal_page(vma, vmf->address, pte);
pte              3701 mm/memory.c    		pte_unmap_unlock(vmf->pte, vmf->ptl);
pte              3707 mm/memory.c    		pte_unmap_unlock(vmf->pte, vmf->ptl);
pte              3719 mm/memory.c    	if (!pte_write(pte))
pte              3733 mm/memory.c    	pte_unmap_unlock(vmf->pte, vmf->ptl);
pte              3832 mm/memory.c    		vmf->pte = NULL;
pte              3843 mm/memory.c    		vmf->pte = pte_offset_map(vmf->pmd, vmf->address);
pte              3844 mm/memory.c    		vmf->orig_pte = *vmf->pte;
pte              3856 mm/memory.c    			pte_unmap(vmf->pte);
pte              3857 mm/memory.c    			vmf->pte = NULL;
pte              3861 mm/memory.c    	if (!vmf->pte) {
pte              3877 mm/memory.c    	if (unlikely(!pte_same(*vmf->pte, entry)))
pte              3885 mm/memory.c    	if (ptep_set_access_flags(vmf->vma, vmf->address, vmf->pte, entry,
pte              3887 mm/memory.c    		update_mmu_cache(vmf->vma, vmf->address, vmf->pte);
pte              3899 mm/memory.c    	pte_unmap_unlock(vmf->pte, vmf->ptl);
pte              4260 mm/memory.c    	pte_t *ptep, pte;
pte              4268 mm/memory.c    	pte = *ptep;
pte              4270 mm/memory.c    	if ((flags & FOLL_WRITE) && !pte_write(pte))
pte              4273 mm/memory.c    	*prot = pgprot_val(pte_pgprot(pte));
pte              4274 mm/memory.c    	*phys = (resource_size_t)pte_pfn(pte) << PAGE_SHIFT;
pte               499 mm/mempolicy.c 	pte_t *pte;
pte               513 mm/mempolicy.c 	pte = pte_offset_map_lock(walk->mm, pmd, addr, &ptl);
pte               514 mm/mempolicy.c 	for (; addr != end; pte++, addr += PAGE_SIZE) {
pte               515 mm/mempolicy.c 		if (!pte_present(*pte))
pte               517 mm/mempolicy.c 		page = vm_normal_page(vma, addr, *pte);
pte               545 mm/mempolicy.c 	pte_unmap_unlock(pte - 1, ptl);
pte               554 mm/mempolicy.c static int queue_pages_hugetlb(pte_t *pte, unsigned long hmask,
pte               565 mm/mempolicy.c 	ptl = huge_pte_lock(hstate_vma(walk->vma), walk->mm, pte);
pte               566 mm/mempolicy.c 	entry = huge_ptep_get(pte);
pte               214 mm/migrate.c   	pte_t pte;
pte               227 mm/migrate.c   		if (!pvmw.pte) {
pte               235 mm/migrate.c   		pte = pte_mkold(mk_pte(new, READ_ONCE(vma->vm_page_prot)));
pte               236 mm/migrate.c   		if (pte_swp_soft_dirty(*pvmw.pte))
pte               237 mm/migrate.c   			pte = pte_mksoft_dirty(pte);
pte               242 mm/migrate.c   		entry = pte_to_swp_entry(*pvmw.pte);
pte               244 mm/migrate.c   			pte = maybe_mkwrite(pte, vma);
pte               248 mm/migrate.c   				entry = make_device_private_entry(new, pte_write(pte));
pte               249 mm/migrate.c   				pte = swp_entry_to_pte(entry);
pte               255 mm/migrate.c   			pte = pte_mkhuge(pte);
pte               256 mm/migrate.c   			pte = arch_make_huge_pte(pte, vma, new, 0);
pte               257 mm/migrate.c   			set_huge_pte_at(vma->vm_mm, pvmw.address, pvmw.pte, pte);
pte               265 mm/migrate.c   			set_pte_at(vma->vm_mm, pvmw.address, pvmw.pte, pte);
pte               279 mm/migrate.c   		update_mmu_cache(vma, pvmw.address, pvmw.pte);
pte               310 mm/migrate.c   	pte_t pte;
pte               315 mm/migrate.c   	pte = *ptep;
pte               316 mm/migrate.c   	if (!is_swap_pte(pte))
pte               319 mm/migrate.c   	entry = pte_to_swp_entry(pte);
pte               348 mm/migrate.c   		struct mm_struct *mm, pte_t *pte)
pte               350 mm/migrate.c   	spinlock_t *ptl = huge_pte_lockptr(hstate_vma(vma), mm, pte);
pte               351 mm/migrate.c   	__migration_entry_wait(mm, pte, ptl);
pte              2250 mm/migrate.c   		pte_t pte;
pte              2252 mm/migrate.c   		pte = *ptep;
pte              2254 mm/migrate.c   		if (pte_none(pte)) {
pte              2260 mm/migrate.c   		if (!pte_present(pte)) {
pte              2268 mm/migrate.c   			entry = pte_to_swp_entry(pte);
pte              2278 mm/migrate.c   			pfn = pte_pfn(pte);
pte              2284 mm/migrate.c   			page = vm_normal_page(migrate->vma, addr, pte);
pte              2286 mm/migrate.c   			mpfn |= pte_write(pte) ? MIGRATE_PFN_WRITE : 0;
pte              2322 mm/migrate.c   			if (pte_soft_dirty(pte))
pte              2334 mm/migrate.c   			if (pte_present(pte))
pte                24 mm/mincore.c   static int mincore_hugetlb(pte_t *pte, unsigned long hmask, unsigned long addr,
pte                35 mm/mincore.c   	present = pte && !huge_pte_none(huge_ptep_get(pte));
pte               145 mm/mincore.c   		pte_t pte = *ptep;
pte               147 mm/mincore.c   		if (pte_none(pte))
pte               150 mm/mincore.c   		else if (pte_present(pte))
pte               153 mm/mincore.c   			swp_entry_t entry = pte_to_swp_entry(pte);
pte               378 mm/mlock.c     	pte_t *pte;
pte               386 mm/mlock.c     	pte = get_locked_pte(vma->vm_mm, start,	&ptl);
pte               397 mm/mlock.c     		pte++;
pte               398 mm/mlock.c     		if (pte_present(*pte))
pte               399 mm/mlock.c     			page = vm_normal_page(vma, start, *pte);
pte               423 mm/mlock.c     	pte_unmap_unlock(pte, ptl);
pte               149 mm/mmu_notifier.c 			       pte_t pte)
pte               157 mm/mmu_notifier.c 			mn->ops->change_pte(mn, mm, address, pte);
pte                42 mm/mprotect.c  	pte_t *pte, oldpte;
pte                61 mm/mprotect.c  	pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl);
pte                71 mm/mprotect.c  		oldpte = *pte;
pte               112 mm/mprotect.c  			oldpte = ptep_modify_prot_start(vma, addr, pte);
pte               123 mm/mprotect.c  			ptep_modify_prot_commit(vma, addr, pte, oldpte, ptent);
pte               138 mm/mprotect.c  				set_pte_at(vma->vm_mm, addr, pte, newpte);
pte               152 mm/mprotect.c  				set_pte_at(vma->vm_mm, addr, pte, newpte);
pte               157 mm/mprotect.c  	} while (pte++, addr += PAGE_SIZE, addr != end);
pte               159 mm/mprotect.c  	pte_unmap_unlock(pte - 1, ptl);
pte               345 mm/mprotect.c  static int prot_none_pte_entry(pte_t *pte, unsigned long addr,
pte               348 mm/mprotect.c  	return pfn_modify_allowed(pte_pfn(*pte), *(pgprot_t *)(walk->private)) ?
pte               352 mm/mprotect.c  static int prot_none_hugetlb_entry(pte_t *pte, unsigned long hmask,
pte               356 mm/mprotect.c  	return pfn_modify_allowed(pte_pfn(*pte), *(pgprot_t *)(walk->private)) ?
pte               100 mm/mremap.c    static pte_t move_soft_dirty_pte(pte_t pte)
pte               107 mm/mremap.c    	if (pte_present(pte))
pte               108 mm/mremap.c    		pte = pte_mksoft_dirty(pte);
pte               109 mm/mremap.c    	else if (is_swap_pte(pte))
pte               110 mm/mremap.c    		pte = pte_swp_mksoft_dirty(pte);
pte               112 mm/mremap.c    	return pte;
pte               121 mm/mremap.c    	pte_t *old_pte, *new_pte, pte;
pte               164 mm/mremap.c    		pte = ptep_get_and_clear(mm, old_addr, old_pte);
pte               176 mm/mremap.c    		if (pte_present(pte))
pte               178 mm/mremap.c    		pte = move_pte(pte, new_vma->vm_page_prot, old_addr, new_addr);
pte               179 mm/mremap.c    		pte = move_soft_dirty_pte(pte);
pte               180 mm/mremap.c    		set_pte_at(mm, new_addr, new_pte, pte);
pte                67 mm/page_idle.c 		if (pvmw.pte) {
pte                72 mm/page_idle.c 			if (ptep_clear_young_notify(vma, addr, pvmw.pte))
pte                18 mm/page_vma_mapped.c 	pvmw->pte = pte_offset_map(pvmw->pmd, pvmw->address);
pte                21 mm/page_vma_mapped.c 			if (!is_swap_pte(*pvmw->pte))
pte                39 mm/page_vma_mapped.c 			if (is_swap_pte(*pvmw->pte)) {
pte                43 mm/page_vma_mapped.c 				entry = pte_to_swp_entry(*pvmw->pte);
pte                46 mm/page_vma_mapped.c 			} else if (!pte_present(*pvmw->pte))
pte                87 mm/page_vma_mapped.c 		if (!is_swap_pte(*pvmw->pte))
pte                89 mm/page_vma_mapped.c 		entry = pte_to_swp_entry(*pvmw->pte);
pte                95 mm/page_vma_mapped.c 	} else if (is_swap_pte(*pvmw->pte)) {
pte                99 mm/page_vma_mapped.c 		entry = pte_to_swp_entry(*pvmw->pte);
pte               105 mm/page_vma_mapped.c 		if (!pte_present(*pvmw->pte))
pte               108 mm/page_vma_mapped.c 		pfn = pte_pfn(*pvmw->pte);
pte               148 mm/page_vma_mapped.c 	if (pvmw->pmd && !pvmw->pte)
pte               151 mm/page_vma_mapped.c 	if (pvmw->pte)
pte               156 mm/page_vma_mapped.c 		pvmw->pte = huge_pte_offset(mm, pvmw->address, page_size(page));
pte               157 mm/page_vma_mapped.c 		if (!pvmw->pte)
pte               160 mm/page_vma_mapped.c 		pvmw->ptl = huge_pte_lockptr(page_hstate(page), mm, pvmw->pte);
pte               230 mm/page_vma_mapped.c 				pte_unmap(pvmw->pte);
pte               237 mm/page_vma_mapped.c 				pvmw->pte++;
pte               239 mm/page_vma_mapped.c 		} while (pte_none(*pvmw->pte));
pte                10 mm/pagewalk.c  	pte_t *pte;
pte                14 mm/pagewalk.c  	pte = pte_offset_map(pmd, addr);
pte                16 mm/pagewalk.c  		err = ops->pte_entry(pte, addr, addr + PAGE_SIZE, walk);
pte                22 mm/pagewalk.c  		pte++;
pte                25 mm/pagewalk.c  	pte_unmap(pte);
pte               190 mm/pagewalk.c  	pte_t *pte;
pte               196 mm/pagewalk.c  		pte = huge_pte_offset(walk->mm, addr & hmask, sz);
pte               198 mm/pagewalk.c  		if (pte)
pte               199 mm/pagewalk.c  			err = ops->hugetlb_entry(pte, hmask, addr, next, walk);
pte                85 mm/pgtable-generic.c 	pte_t pte;
pte                86 mm/pgtable-generic.c 	pte = ptep_get_and_clear(mm, address, ptep);
pte                87 mm/pgtable-generic.c 	if (pte_accessible(mm, pte))
pte                89 mm/pgtable-generic.c 	return pte;
pte               774 mm/rmap.c      		if (pvmw.pte) {
pte               776 mm/rmap.c      						pvmw.pte)) {
pte               909 mm/rmap.c      		if (pvmw.pte) {
pte               911 mm/rmap.c      			pte_t *pte = pvmw.pte;
pte               913 mm/rmap.c      			if (!pte_dirty(*pte) && !pte_write(*pte))
pte               916 mm/rmap.c      			flush_cache_page(vma, address, pte_pfn(*pte));
pte               917 mm/rmap.c      			entry = ptep_clear_flush(vma, address, pte);
pte               920 mm/rmap.c      			set_pte_at(vma->vm_mm, address, pte, entry);
pte              1393 mm/rmap.c      		if (!pvmw.pte && (flags & TTU_MIGRATION)) {
pte              1425 mm/rmap.c      		VM_BUG_ON_PAGE(!pvmw.pte, page);
pte              1427 mm/rmap.c      		subpage = page - page_to_pfn(page) + pte_pfn(*pvmw.pte);
pte              1431 mm/rmap.c      			if (huge_pmd_unshare(mm, &address, pvmw.pte)) {
pte              1464 mm/rmap.c      			pteval = ptep_get_and_clear(mm, pvmw.address, pvmw.pte);
pte              1475 mm/rmap.c      			set_pte_at(mm, pvmw.address, pvmw.pte, swp_pte);
pte              1493 mm/rmap.c      						pvmw.pte)) {
pte              1501 mm/rmap.c      		flush_cache_page(vma, address, pte_pfn(*pvmw.pte));
pte              1511 mm/rmap.c      			pteval = ptep_get_and_clear(mm, address, pvmw.pte);
pte              1515 mm/rmap.c      			pteval = ptep_clear_flush(vma, address, pvmw.pte);
pte              1530 mm/rmap.c      						     pvmw.pte, pteval,
pte              1534 mm/rmap.c      				set_pte_at(mm, address, pvmw.pte, pteval);
pte              1558 mm/rmap.c      				set_pte_at(mm, address, pvmw.pte, pteval);
pte              1574 mm/rmap.c      			set_pte_at(mm, address, pvmw.pte, swp_pte);
pte              1610 mm/rmap.c      				set_pte_at(mm, address, pvmw.pte, pteval);
pte              1618 mm/rmap.c      				set_pte_at(mm, address, pvmw.pte, pteval);
pte              1624 mm/rmap.c      				set_pte_at(mm, address, pvmw.pte, pteval);
pte              1640 mm/rmap.c      			set_pte_at(mm, address, pvmw.pte, swp_pte);
pte               132 mm/sparse-vmemmap.c void __meminit vmemmap_verify(pte_t *pte, int node,
pte               135 mm/sparse-vmemmap.c 	unsigned long pfn = pte_pfn(*pte);
pte               145 mm/sparse-vmemmap.c 	pte_t *pte = pte_offset_kernel(pmd, addr);
pte               146 mm/sparse-vmemmap.c 	if (pte_none(*pte)) {
pte               152 mm/sparse-vmemmap.c 		set_pte_at(&init_mm, addr, pte, entry);
pte               154 mm/sparse-vmemmap.c 	return pte;
pte               224 mm/sparse-vmemmap.c 	pte_t *pte;
pte               239 mm/sparse-vmemmap.c 		pte = vmemmap_pte_populate(pmd, addr, node);
pte               240 mm/sparse-vmemmap.c 		if (!pte)
pte               242 mm/sparse-vmemmap.c 		vmemmap_verify(pte, node, addr, addr + PAGE_SIZE);
pte               648 mm/swap_state.c 	pte_t *pte, *orig_pte;
pte               662 mm/swap_state.c 	orig_pte = pte = pte_offset_map(vmf->pmd, faddr);
pte               663 mm/swap_state.c 	entry = pte_to_swp_entry(*pte);
pte               697 mm/swap_state.c 	pte -= ra_info->offset;
pte               699 mm/swap_state.c 	ra_info->ptes = pte;
pte               703 mm/swap_state.c 		*tpte++ = *pte++;
pte               728 mm/swap_state.c 	pte_t *pte, pentry;
pte               739 mm/swap_state.c 	for (i = 0, pte = ra_info.ptes; i < ra_info.nr_pte;
pte               740 mm/swap_state.c 	     i++, pte++) {
pte               741 mm/swap_state.c 		pentry = *pte;
pte              1843 mm/swapfile.c  static inline int pte_same_as_swp(pte_t pte, pte_t swp_pte)
pte              1845 mm/swapfile.c  	return pte_same(pte_swp_clear_soft_dirty(pte), swp_pte);
pte              1859 mm/swapfile.c  	pte_t *pte;
pte              1873 mm/swapfile.c  	pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl);
pte              1874 mm/swapfile.c  	if (unlikely(!pte_same_as_swp(*pte, swp_entry_to_pte(entry)))) {
pte              1883 mm/swapfile.c  	set_pte_at(vma->vm_mm, addr, pte,
pte              1900 mm/swapfile.c  	pte_unmap_unlock(pte, ptl);
pte              1916 mm/swapfile.c  	pte_t *pte;
pte              1923 mm/swapfile.c  	pte = pte_offset_map(pmd, addr);
pte              1927 mm/swapfile.c  		if (!is_swap_pte(*pte))
pte              1930 mm/swapfile.c  		entry = pte_to_swp_entry(*pte);
pte              1938 mm/swapfile.c  		pte_unmap(pte);
pte              1968 mm/swapfile.c  		pte = pte_offset_map(pmd, addr);
pte              1969 mm/swapfile.c  	} while (pte++, addr += PAGE_SIZE, addr != end);
pte              1970 mm/swapfile.c  	pte_unmap(pte - 1);
pte                66 mm/vmalloc.c   	pte_t *pte;
pte                68 mm/vmalloc.c   	pte = pte_offset_kernel(pmd, addr);
pte                70 mm/vmalloc.c   		pte_t ptent = ptep_get_and_clear(&init_mm, addr, pte);
pte                72 mm/vmalloc.c   	} while (pte++, addr += PAGE_SIZE, addr != end);
pte               141 mm/vmalloc.c   	pte_t *pte;
pte               148 mm/vmalloc.c   	pte = pte_alloc_kernel(pmd, addr);
pte               149 mm/vmalloc.c   	if (!pte)
pte               154 mm/vmalloc.c   		if (WARN_ON(!pte_none(*pte)))
pte               158 mm/vmalloc.c   		set_pte_at(&init_mm, addr, pte, mk_pte(page, prot));
pte               160 mm/vmalloc.c   	} while (pte++, addr += PAGE_SIZE, addr != end);
pte               278 mm/vmalloc.c   	pte_t *ptep, pte;
pte               310 mm/vmalloc.c   	pte = *ptep;
pte               311 mm/vmalloc.c   	if (pte_present(pte))
pte               312 mm/vmalloc.c   		page = pte_page(pte);
pte              3077 mm/vmalloc.c   static int f(pte_t *pte, unsigned long addr, void *data)
pte              3082 mm/vmalloc.c   		*(*p) = pte;
pte               444 tools/lib/traceevent/plugins/plugin_kvm.c 	unsigned long pte = args[0];
pte               445 tools/lib/traceevent/plugins/plugin_kvm.c 	return pte & PT_WRITABLE_MASK;
pte               191 tools/testing/selftests/kvm/lib/aarch64/processor.c 	uint64_t pte, *ptep;
pte               196 tools/testing/selftests/kvm/lib/aarch64/processor.c 	for (pte = page; pte < page + ptrs_per_pte(vm) * 8; pte += 8) {
pte               197 tools/testing/selftests/kvm/lib/aarch64/processor.c 		ptep = addr_gpa2hva(vm, pte);
pte               200 tools/testing/selftests/kvm/lib/aarch64/processor.c 		printf("%*s%s: %lx: %lx at %p\n", indent, "", type[level], pte, *ptep, ptep);
pte               158 tools/testing/selftests/kvm/lib/s390x/processor.c 	uint64_t *pte, ptea;
pte               161 tools/testing/selftests/kvm/lib/s390x/processor.c 		pte = addr_gpa2hva(vm, ptea);
pte               162 tools/testing/selftests/kvm/lib/s390x/processor.c 		if (*pte & PAGE_INVALID)
pte               165 tools/testing/selftests/kvm/lib/s390x/processor.c 			indent, "", ptea, *pte);
pte               322 tools/testing/selftests/kvm/lib/x86_64/processor.c 	struct pageTableEntry *pte;
pte               323 tools/testing/selftests/kvm/lib/x86_64/processor.c 	pte = addr_gpa2hva(vm, pde[index[1]].address * vm->page_size);
pte               324 tools/testing/selftests/kvm/lib/x86_64/processor.c 	pte[index[0]].address = paddr >> vm->page_shift;
pte               325 tools/testing/selftests/kvm/lib/x86_64/processor.c 	pte[index[0]].writable = true;
pte               326 tools/testing/selftests/kvm/lib/x86_64/processor.c 	pte[index[0]].present = 1;
pte               348 tools/testing/selftests/kvm/lib/x86_64/processor.c 	struct pageTableEntry *pte, *pte_start;
pte               401 tools/testing/selftests/kvm/lib/x86_64/processor.c 					pte = &pte_start[n4];
pte               402 tools/testing/selftests/kvm/lib/x86_64/processor.c 					if (!pte->present)
pte               408 tools/testing/selftests/kvm/lib/x86_64/processor.c 						pte - pte_start, pte,
pte               409 tools/testing/selftests/kvm/lib/x86_64/processor.c 						addr_hva2gpa(vm, pte),
pte               410 tools/testing/selftests/kvm/lib/x86_64/processor.c 						(uint64_t) pte->address,
pte               411 tools/testing/selftests/kvm/lib/x86_64/processor.c 						pte->writable,
pte               412 tools/testing/selftests/kvm/lib/x86_64/processor.c 						pte->execute_disable,
pte               413 tools/testing/selftests/kvm/lib/x86_64/processor.c 						pte->dirty,
pte               548 tools/testing/selftests/kvm/lib/x86_64/processor.c 	struct pageTableEntry *pte;
pte               572 tools/testing/selftests/kvm/lib/x86_64/processor.c 	pte = addr_gpa2hva(vm, pde[index[1]].address * vm->page_size);
pte               573 tools/testing/selftests/kvm/lib/x86_64/processor.c 	if (!pte[index[0]].present)
pte               576 tools/testing/selftests/kvm/lib/x86_64/processor.c 	return (pte[index[0]].address * vm->page_size) + (gva & 0xfffu);
pte               459 tools/testing/selftests/kvm/lib/x86_64/vmx.c 	struct eptPageTableEntry *pte;
pte               460 tools/testing/selftests/kvm/lib/x86_64/vmx.c 	pte = addr_gpa2hva(vm, pde[index[1]].address * vm->page_size);
pte               461 tools/testing/selftests/kvm/lib/x86_64/vmx.c 	pte[index[0]].address = paddr >> vm->page_shift;
pte               462 tools/testing/selftests/kvm/lib/x86_64/vmx.c 	pte[index[0]].writable = true;
pte               463 tools/testing/selftests/kvm/lib/x86_64/vmx.c 	pte[index[0]].readable = true;
pte               464 tools/testing/selftests/kvm/lib/x86_64/vmx.c 	pte[index[0]].executable = true;
pte               470 tools/testing/selftests/kvm/lib/x86_64/vmx.c 	pte[index[0]].accessed = true;
pte               471 tools/testing/selftests/kvm/lib/x86_64/vmx.c 	pte[index[0]].dirty = true;
pte                72 virt/kvm/arm/mmu.c static void kvm_flush_dcache_pte(pte_t pte)
pte                74 virt/kvm/arm/mmu.c 	__kvm_flush_dcache_pte(pte);
pte               246 virt/kvm/arm/mmu.c 	pte_t *pte, *start_pte;
pte               248 virt/kvm/arm/mmu.c 	start_pte = pte = pte_offset_kernel(pmd, addr);
pte               250 virt/kvm/arm/mmu.c 		if (!pte_none(*pte)) {
pte               251 virt/kvm/arm/mmu.c 			pte_t old_pte = *pte;
pte               253 virt/kvm/arm/mmu.c 			kvm_set_pte(pte, __pte(0));
pte               260 virt/kvm/arm/mmu.c 			put_page(virt_to_page(pte));
pte               262 virt/kvm/arm/mmu.c 	} while (pte++, addr += PAGE_SIZE, addr != end);
pte               368 virt/kvm/arm/mmu.c 	pte_t *pte;
pte               370 virt/kvm/arm/mmu.c 	pte = pte_offset_kernel(pmd, addr);
pte               372 virt/kvm/arm/mmu.c 		if (!pte_none(*pte) && !kvm_is_device_pfn(pte_pfn(*pte)))
pte               373 virt/kvm/arm/mmu.c 			kvm_flush_dcache_pte(*pte);
pte               374 virt/kvm/arm/mmu.c 	} while (pte++, addr += PAGE_SIZE, addr != end);
pte               481 virt/kvm/arm/mmu.c 	pte_t *pte, *start_pte;
pte               483 virt/kvm/arm/mmu.c 	start_pte = pte = pte_offset_kernel(pmd, addr);
pte               485 virt/kvm/arm/mmu.c 		if (!pte_none(*pte)) {
pte               486 virt/kvm/arm/mmu.c 			kvm_set_pte(pte, __pte(0));
pte               487 virt/kvm/arm/mmu.c 			put_page(virt_to_page(pte));
pte               489 virt/kvm/arm/mmu.c 	} while (pte++, addr += PAGE_SIZE, addr != end);
pte               614 virt/kvm/arm/mmu.c 	pte_t *pte;
pte               619 virt/kvm/arm/mmu.c 		pte = pte_offset_kernel(pmd, addr);
pte               620 virt/kvm/arm/mmu.c 		kvm_set_pte(pte, kvm_pfn_pte(pfn, prot));
pte               621 virt/kvm/arm/mmu.c 		get_page(virt_to_page(pte));
pte               631 virt/kvm/arm/mmu.c 	pte_t *pte;
pte               641 virt/kvm/arm/mmu.c 			pte = pte_alloc_one_kernel(NULL);
pte               642 virt/kvm/arm/mmu.c 			if (!pte) {
pte               646 virt/kvm/arm/mmu.c 			kvm_pmd_populate(pmd, pte);
pte              1227 virt/kvm/arm/mmu.c 	pte_t *pte, old_pte;
pte              1278 virt/kvm/arm/mmu.c 		pte = mmu_memory_cache_alloc(cache);
pte              1279 virt/kvm/arm/mmu.c 		kvm_pmd_populate(pmd, pte);
pte              1283 virt/kvm/arm/mmu.c 	pte = pte_offset_kernel(pmd, addr);
pte              1285 virt/kvm/arm/mmu.c 	if (iomap && pte_present(*pte))
pte              1289 virt/kvm/arm/mmu.c 	old_pte = *pte;
pte              1295 virt/kvm/arm/mmu.c 		kvm_set_pte(pte, __pte(0));
pte              1298 virt/kvm/arm/mmu.c 		get_page(virt_to_page(pte));
pte              1301 virt/kvm/arm/mmu.c 	kvm_set_pte(pte, *new_pte);
pte              1306 virt/kvm/arm/mmu.c static int stage2_ptep_test_and_clear_young(pte_t *pte)
pte              1308 virt/kvm/arm/mmu.c 	if (pte_young(*pte)) {
pte              1309 virt/kvm/arm/mmu.c 		*pte = pte_mkold(*pte);
pte              1315 virt/kvm/arm/mmu.c static int stage2_ptep_test_and_clear_young(pte_t *pte)
pte              1317 virt/kvm/arm/mmu.c 	return __ptep_test_and_clear_young(pte);
pte              1351 virt/kvm/arm/mmu.c 		pte_t pte = kvm_pfn_pte(pfn, PAGE_S2_DEVICE);
pte              1354 virt/kvm/arm/mmu.c 			pte = kvm_s2pte_mkwrite(pte);
pte              1362 virt/kvm/arm/mmu.c 		ret = stage2_set_pte(kvm, &cache, addr, &pte,
pte              1431 virt/kvm/arm/mmu.c 	pte_t *pte;
pte              1433 virt/kvm/arm/mmu.c 	pte = pte_offset_kernel(pmd, addr);
pte              1435 virt/kvm/arm/mmu.c 		if (!pte_none(*pte)) {
pte              1436 virt/kvm/arm/mmu.c 			if (!kvm_s2pte_readonly(pte))
pte              1437 virt/kvm/arm/mmu.c 				kvm_set_s2pte_readonly(pte);
pte              1439 virt/kvm/arm/mmu.c 	} while (pte++, addr += PAGE_SIZE, addr != end);
pte              1865 virt/kvm/arm/mmu.c 	pte_t *pte;
pte              1873 virt/kvm/arm/mmu.c 	if (!stage2_get_leaf_entry(vcpu->kvm, fault_ipa, &pud, &pmd, &pte))
pte              1885 virt/kvm/arm/mmu.c 		*pte = pte_mkyoung(*pte);	/* Just a page... */
pte              1886 virt/kvm/arm/mmu.c 		pfn = pte_pfn(*pte);
pte              2064 virt/kvm/arm/mmu.c 	pte_t *pte = (pte_t *)data;
pte              2074 virt/kvm/arm/mmu.c 	stage2_set_pte(kvm, NULL, gpa, pte, 0);
pte              2079 virt/kvm/arm/mmu.c int kvm_set_spte_hva(struct kvm *kvm, unsigned long hva, pte_t pte)
pte              2082 virt/kvm/arm/mmu.c 	kvm_pfn_t pfn = pte_pfn(pte);
pte              2105 virt/kvm/arm/mmu.c 	pte_t *pte;
pte              2108 virt/kvm/arm/mmu.c 	if (!stage2_get_leaf_entry(kvm, gpa, &pud, &pmd, &pte))
pte              2116 virt/kvm/arm/mmu.c 		return stage2_ptep_test_and_clear_young(pte);
pte              2123 virt/kvm/arm/mmu.c 	pte_t *pte;
pte              2126 virt/kvm/arm/mmu.c 	if (!stage2_get_leaf_entry(kvm, gpa, &pud, &pmd, &pte))
pte              2134 virt/kvm/arm/mmu.c 		return pte_young(*pte);
pte               398 virt/kvm/kvm_main.c 					pte_t pte)
pte               407 virt/kvm/kvm_main.c 	if (kvm_set_spte_hva(kvm, address, pte))