gpte              188 arch/powerpc/include/asm/kvm_book3s.h 				      struct kvmppc_pte *gpte, u64 root,
gpte              191 arch/powerpc/include/asm/kvm_book3s.h 			struct kvmppc_pte *gpte, u64 table,
gpte              194 arch/powerpc/include/asm/kvm_book3s.h 			struct kvmppc_pte *gpte, bool data, bool iswrite);
gpte              206 arch/powerpc/kvm/book3s_64_mmu.c 				      struct kvmppc_pte *gpte, bool data,
gpte              227 arch/powerpc/kvm/book3s_64_mmu.c 		gpte->eaddr = eaddr;
gpte              228 arch/powerpc/kvm/book3s_64_mmu.c 		gpte->vpage = kvmppc_mmu_book3s_64_ea_to_vp(vcpu, eaddr, data);
gpte              229 arch/powerpc/kvm/book3s_64_mmu.c 		gpte->raddr = vcpu->arch.magic_page_pa | (gpte->raddr & 0xfff);
gpte              230 arch/powerpc/kvm/book3s_64_mmu.c 		gpte->raddr &= KVM_PAM;
gpte              231 arch/powerpc/kvm/book3s_64_mmu.c 		gpte->may_execute = true;
gpte              232 arch/powerpc/kvm/book3s_64_mmu.c 		gpte->may_read = true;
gpte              233 arch/powerpc/kvm/book3s_64_mmu.c 		gpte->may_write = true;
gpte              234 arch/powerpc/kvm/book3s_64_mmu.c 		gpte->page_size = MMU_PAGE_4K;
gpte              235 arch/powerpc/kvm/book3s_64_mmu.c 		gpte->wimg = HPTE_R_M;
gpte              308 arch/powerpc/kvm/book3s_64_mmu.c 	gpte->eaddr = eaddr;
gpte              309 arch/powerpc/kvm/book3s_64_mmu.c 	gpte->vpage = kvmppc_mmu_book3s_64_ea_to_vp(vcpu, eaddr, data);
gpte              312 arch/powerpc/kvm/book3s_64_mmu.c 	gpte->raddr = (r & HPTE_R_RPN & ~eaddr_mask) | (eaddr & eaddr_mask);
gpte              313 arch/powerpc/kvm/book3s_64_mmu.c 	gpte->page_size = pgsize;
gpte              314 arch/powerpc/kvm/book3s_64_mmu.c 	gpte->may_execute = ((r & HPTE_R_N) ? false : true);
gpte              317 arch/powerpc/kvm/book3s_64_mmu.c 		gpte->may_execute = true;
gpte              318 arch/powerpc/kvm/book3s_64_mmu.c 	gpte->may_read = false;
gpte              319 arch/powerpc/kvm/book3s_64_mmu.c 	gpte->may_write = false;
gpte              320 arch/powerpc/kvm/book3s_64_mmu.c 	gpte->wimg = r & HPTE_R_WIMG;
gpte              327 arch/powerpc/kvm/book3s_64_mmu.c 		gpte->may_write = true;
gpte              333 arch/powerpc/kvm/book3s_64_mmu.c 		gpte->may_read = true;
gpte              339 arch/powerpc/kvm/book3s_64_mmu.c 		eaddr, avpn, gpte->vpage, gpte->raddr);
gpte              343 arch/powerpc/kvm/book3s_64_mmu.c 	if (gpte->may_read && !(r & HPTE_R_R)) {
gpte              355 arch/powerpc/kvm/book3s_64_mmu.c 	if (iswrite && gpte->may_write && !(r & HPTE_R_C)) {
gpte              365 arch/powerpc/kvm/book3s_64_mmu.c 	if (!gpte->may_read || (iswrite && !gpte->may_write))
gpte              341 arch/powerpc/kvm/book3s_64_mmu_hv.c 			struct kvmppc_pte *gpte, bool data, bool iswrite)
gpte              353 arch/powerpc/kvm/book3s_64_mmu_hv.c 		return kvmppc_mmu_radix_xlate(vcpu, eaddr, gpte, data, iswrite);
gpte              383 arch/powerpc/kvm/book3s_64_mmu_hv.c 	gpte->eaddr = eaddr;
gpte              384 arch/powerpc/kvm/book3s_64_mmu_hv.c 	gpte->vpage = ((v & HPTE_V_AVPN) << 4) | ((eaddr >> 12) & 0xfff);
gpte              392 arch/powerpc/kvm/book3s_64_mmu_hv.c 	gpte->may_read = hpte_read_permission(pp, key);
gpte              393 arch/powerpc/kvm/book3s_64_mmu_hv.c 	gpte->may_write = hpte_write_permission(pp, key);
gpte              394 arch/powerpc/kvm/book3s_64_mmu_hv.c 	gpte->may_execute = gpte->may_read && !(gr & (HPTE_R_N | HPTE_R_G));
gpte              400 arch/powerpc/kvm/book3s_64_mmu_hv.c 			gpte->may_read = 0;
gpte              402 arch/powerpc/kvm/book3s_64_mmu_hv.c 			gpte->may_write = 0;
gpte              406 arch/powerpc/kvm/book3s_64_mmu_hv.c 	gpte->raddr = kvmppc_mmu_get_real_addr(v, gr, eaddr);
gpte              128 arch/powerpc/kvm/book3s_64_mmu_radix.c 			       struct kvmppc_pte *gpte, u64 root,
gpte              193 arch/powerpc/kvm/book3s_64_mmu_radix.c 	gpte->page_size = ps;
gpte              194 arch/powerpc/kvm/book3s_64_mmu_radix.c 	gpte->page_shift = offset;
gpte              196 arch/powerpc/kvm/book3s_64_mmu_radix.c 	gpte->eaddr = eaddr;
gpte              197 arch/powerpc/kvm/book3s_64_mmu_radix.c 	gpte->raddr = gpa;
gpte              200 arch/powerpc/kvm/book3s_64_mmu_radix.c 	gpte->may_read = !!(pte & _PAGE_READ);
gpte              201 arch/powerpc/kvm/book3s_64_mmu_radix.c 	gpte->may_write = !!(pte & _PAGE_WRITE);
gpte              202 arch/powerpc/kvm/book3s_64_mmu_radix.c 	gpte->may_execute = !!(pte & _PAGE_EXEC);
gpte              204 arch/powerpc/kvm/book3s_64_mmu_radix.c 	gpte->rc = pte & (_PAGE_ACCESSED | _PAGE_DIRTY);
gpte              221 arch/powerpc/kvm/book3s_64_mmu_radix.c 				     struct kvmppc_pte *gpte, u64 table,
gpte              246 arch/powerpc/kvm/book3s_64_mmu_radix.c 	return kvmppc_mmu_walk_radix_tree(vcpu, eaddr, gpte, root, pte_ret_p);
gpte              250 arch/powerpc/kvm/book3s_64_mmu_radix.c 			   struct kvmppc_pte *gpte, bool data, bool iswrite)
gpte              268 arch/powerpc/kvm/book3s_64_mmu_radix.c 	ret = kvmppc_mmu_radix_translate_table(vcpu, eaddr, gpte,
gpte              276 arch/powerpc/kvm/book3s_64_mmu_radix.c 			gpte->may_read = 0;
gpte              277 arch/powerpc/kvm/book3s_64_mmu_radix.c 			gpte->may_write = 0;
gpte              278 arch/powerpc/kvm/book3s_64_mmu_radix.c 			gpte->may_execute = 0;
gpte              284 arch/powerpc/kvm/book3s_64_mmu_radix.c 				gpte->may_read = 0;
gpte              286 arch/powerpc/kvm/book3s_64_mmu_radix.c 				gpte->may_write = 0;
gpte              288 arch/powerpc/kvm/book3s_64_mmu_radix.c 				gpte->may_execute = 0;
gpte             1198 arch/powerpc/kvm/book3s_hv_nested.c 				       struct kvmppc_pte gpte,
gpte             1210 arch/powerpc/kvm/book3s_hv_nested.c 	if (pgflags & ~gpte.rc)
gpte             1216 arch/powerpc/kvm/book3s_hv_nested.c 				     gpte.raddr, kvm->arch.lpid);
gpte             1267 arch/powerpc/kvm/book3s_hv_nested.c 	struct kvmppc_pte gpte;
gpte             1290 arch/powerpc/kvm/book3s_hv_nested.c 	ret = kvmhv_translate_addr_nested(vcpu, gp, n_gpa, dsisr, &gpte);
gpte             1306 arch/powerpc/kvm/book3s_hv_nested.c 		ret = kvmhv_handle_nested_set_rc(vcpu, gp, n_gpa, gpte, dsisr);
gpte             1323 arch/powerpc/kvm/book3s_hv_nested.c 	l1_shift = gpte.page_shift;
gpte             1330 arch/powerpc/kvm/book3s_hv_nested.c 	gpa = gpte.raddr;
gpte             1388 arch/powerpc/kvm/book3s_hv_nested.c 	perm |= gpte.may_read ? 0UL : _PAGE_READ;
gpte             1389 arch/powerpc/kvm/book3s_hv_nested.c 	perm |= gpte.may_write ? 0UL : _PAGE_WRITE;
gpte             1390 arch/powerpc/kvm/book3s_hv_nested.c 	perm |= gpte.may_execute ? 0UL : _PAGE_EXEC;
gpte             1392 arch/powerpc/kvm/book3s_hv_nested.c 	perm |= (gpte.rc & _PAGE_ACCESSED) ? 0UL : _PAGE_ACCESSED;
gpte             1393 arch/powerpc/kvm/book3s_hv_nested.c 	perm |= ((gpte.rc & _PAGE_DIRTY) && writing) ? 0UL : _PAGE_DIRTY;
gpte              638 arch/x86/kvm/mmu.c static gfn_t pse36_gfn_delta(u32 gpte)
gpte              642 arch/x86/kvm/mmu.c 	return (gpte & PT32_DIR_PSE36_MASK) << shift;
gpte             4016 arch/x86/kvm/mmu.c static bool is_rsvd_bits_set(struct kvm_mmu *mmu, u64 gpte, int level)
gpte             4018 arch/x86/kvm/mmu.c 	return __is_rsvd_bits_set(&mmu->guest_rsvd_check, gpte, level);
gpte             4481 arch/x86/kvm/mmu.c 				unsigned level, unsigned gpte)
gpte             4488 arch/x86/kvm/mmu.c 	gpte &= level - mmu->last_nonleaf_level;
gpte             4495 arch/x86/kvm/mmu.c 	gpte |= level - PT_PAGE_TABLE_LEVEL - 1;
gpte             4497 arch/x86/kvm/mmu.c 	return gpte & PT_PAGE_SIZE_MASK;
gpte               99 arch/x86/kvm/paging_tmpl.h static gfn_t gpte_to_gfn_lvl(pt_element_t gpte, int lvl)
gpte              101 arch/x86/kvm/paging_tmpl.h 	return (gpte & PT_LVL_ADDR_MASK(lvl)) >> PAGE_SHIFT;
gpte              105 arch/x86/kvm/paging_tmpl.h 					     unsigned gpte)
gpte              117 arch/x86/kvm/paging_tmpl.h 	mask |= (gpte >> (PT_GUEST_DIRTY_SHIFT - PT_WRITABLE_SHIFT)) &
gpte              176 arch/x86/kvm/paging_tmpl.h 				  u64 gpte)
gpte              178 arch/x86/kvm/paging_tmpl.h 	if (is_rsvd_bits_set(vcpu->arch.mmu, gpte, PT_PAGE_TABLE_LEVEL))
gpte              181 arch/x86/kvm/paging_tmpl.h 	if (!FNAME(is_present_gpte)(gpte))
gpte              186 arch/x86/kvm/paging_tmpl.h 	    !(gpte & PT_GUEST_ACCESSED_MASK))
gpte              202 arch/x86/kvm/paging_tmpl.h static inline unsigned FNAME(gpte_access)(u64 gpte)
gpte              206 arch/x86/kvm/paging_tmpl.h 	access = ((gpte & VMX_EPT_WRITABLE_MASK) ? ACC_WRITE_MASK : 0) |
gpte              207 arch/x86/kvm/paging_tmpl.h 		((gpte & VMX_EPT_EXECUTABLE_MASK) ? ACC_EXEC_MASK : 0) |
gpte              208 arch/x86/kvm/paging_tmpl.h 		((gpte & VMX_EPT_READABLE_MASK) ? ACC_USER_MASK : 0);
gpte              212 arch/x86/kvm/paging_tmpl.h 	access = gpte & (PT_WRITABLE_MASK | PT_USER_MASK | PT_PRESENT_MASK);
gpte              214 arch/x86/kvm/paging_tmpl.h 	access ^= (gpte >> PT64_NX_SHIFT);
gpte              282 arch/x86/kvm/paging_tmpl.h static inline unsigned FNAME(gpte_pkeys)(struct kvm_vcpu *vcpu, u64 gpte)
gpte              286 arch/x86/kvm/paging_tmpl.h 	pte_t pte = {.pte = gpte};
gpte              517 arch/x86/kvm/paging_tmpl.h 		     u64 *spte, pt_element_t gpte, bool no_dirty_log)
gpte              523 arch/x86/kvm/paging_tmpl.h 	if (FNAME(prefetch_invalid_gpte)(vcpu, sp, spte, gpte))
gpte              526 arch/x86/kvm/paging_tmpl.h 	pgprintk("%s: gpte %llx spte %p\n", __func__, (u64)gpte, spte);
gpte              528 arch/x86/kvm/paging_tmpl.h 	gfn = gpte_to_gfn(gpte);
gpte              529 arch/x86/kvm/paging_tmpl.h 	pte_access = sp->role.access & FNAME(gpte_access)(gpte);
gpte              530 arch/x86/kvm/paging_tmpl.h 	FNAME(protect_clean_gpte)(vcpu->arch.mmu, &pte_access, gpte);
gpte              550 arch/x86/kvm/paging_tmpl.h 	pt_element_t gpte = *(const pt_element_t *)pte;
gpte              552 arch/x86/kvm/paging_tmpl.h 	FNAME(prefetch_gpte)(vcpu, sp, spte, gpte, false);
gpte              919 arch/x86/kvm/paging_tmpl.h 			pt_element_t gpte;
gpte              935 arch/x86/kvm/paging_tmpl.h 			if (kvm_vcpu_read_guest_atomic(vcpu, pte_gpa, &gpte,
gpte              939 arch/x86/kvm/paging_tmpl.h 			FNAME(update_pte)(vcpu, sp, sptep, &gpte);
gpte             1021 arch/x86/kvm/paging_tmpl.h 		pt_element_t gpte;
gpte             1030 arch/x86/kvm/paging_tmpl.h 		if (kvm_vcpu_read_guest_atomic(vcpu, pte_gpa, &gpte,
gpte             1034 arch/x86/kvm/paging_tmpl.h 		if (FNAME(prefetch_invalid_gpte)(vcpu, sp, &sp->spt[i], gpte)) {
gpte             1045 arch/x86/kvm/paging_tmpl.h 		gfn = gpte_to_gfn(gpte);
gpte             1047 arch/x86/kvm/paging_tmpl.h 		pte_access &= FNAME(gpte_access)(gpte);
gpte             1048 arch/x86/kvm/paging_tmpl.h 		FNAME(protect_clean_gpte)(vcpu->arch.mmu, &pte_access, gpte);