gpte 188 arch/powerpc/include/asm/kvm_book3s.h struct kvmppc_pte *gpte, u64 root, gpte 191 arch/powerpc/include/asm/kvm_book3s.h struct kvmppc_pte *gpte, u64 table, gpte 194 arch/powerpc/include/asm/kvm_book3s.h struct kvmppc_pte *gpte, bool data, bool iswrite); gpte 206 arch/powerpc/kvm/book3s_64_mmu.c struct kvmppc_pte *gpte, bool data, gpte 227 arch/powerpc/kvm/book3s_64_mmu.c gpte->eaddr = eaddr; gpte 228 arch/powerpc/kvm/book3s_64_mmu.c gpte->vpage = kvmppc_mmu_book3s_64_ea_to_vp(vcpu, eaddr, data); gpte 229 arch/powerpc/kvm/book3s_64_mmu.c gpte->raddr = vcpu->arch.magic_page_pa | (gpte->raddr & 0xfff); gpte 230 arch/powerpc/kvm/book3s_64_mmu.c gpte->raddr &= KVM_PAM; gpte 231 arch/powerpc/kvm/book3s_64_mmu.c gpte->may_execute = true; gpte 232 arch/powerpc/kvm/book3s_64_mmu.c gpte->may_read = true; gpte 233 arch/powerpc/kvm/book3s_64_mmu.c gpte->may_write = true; gpte 234 arch/powerpc/kvm/book3s_64_mmu.c gpte->page_size = MMU_PAGE_4K; gpte 235 arch/powerpc/kvm/book3s_64_mmu.c gpte->wimg = HPTE_R_M; gpte 308 arch/powerpc/kvm/book3s_64_mmu.c gpte->eaddr = eaddr; gpte 309 arch/powerpc/kvm/book3s_64_mmu.c gpte->vpage = kvmppc_mmu_book3s_64_ea_to_vp(vcpu, eaddr, data); gpte 312 arch/powerpc/kvm/book3s_64_mmu.c gpte->raddr = (r & HPTE_R_RPN & ~eaddr_mask) | (eaddr & eaddr_mask); gpte 313 arch/powerpc/kvm/book3s_64_mmu.c gpte->page_size = pgsize; gpte 314 arch/powerpc/kvm/book3s_64_mmu.c gpte->may_execute = ((r & HPTE_R_N) ? false : true); gpte 317 arch/powerpc/kvm/book3s_64_mmu.c gpte->may_execute = true; gpte 318 arch/powerpc/kvm/book3s_64_mmu.c gpte->may_read = false; gpte 319 arch/powerpc/kvm/book3s_64_mmu.c gpte->may_write = false; gpte 320 arch/powerpc/kvm/book3s_64_mmu.c gpte->wimg = r & HPTE_R_WIMG; gpte 327 arch/powerpc/kvm/book3s_64_mmu.c gpte->may_write = true; gpte 333 arch/powerpc/kvm/book3s_64_mmu.c gpte->may_read = true; gpte 339 arch/powerpc/kvm/book3s_64_mmu.c eaddr, avpn, gpte->vpage, gpte->raddr); gpte 343 arch/powerpc/kvm/book3s_64_mmu.c if (gpte->may_read && !(r & HPTE_R_R)) { gpte 355 arch/powerpc/kvm/book3s_64_mmu.c if (iswrite && gpte->may_write && !(r & HPTE_R_C)) { gpte 365 arch/powerpc/kvm/book3s_64_mmu.c if (!gpte->may_read || (iswrite && !gpte->may_write)) gpte 341 arch/powerpc/kvm/book3s_64_mmu_hv.c struct kvmppc_pte *gpte, bool data, bool iswrite) gpte 353 arch/powerpc/kvm/book3s_64_mmu_hv.c return kvmppc_mmu_radix_xlate(vcpu, eaddr, gpte, data, iswrite); gpte 383 arch/powerpc/kvm/book3s_64_mmu_hv.c gpte->eaddr = eaddr; gpte 384 arch/powerpc/kvm/book3s_64_mmu_hv.c gpte->vpage = ((v & HPTE_V_AVPN) << 4) | ((eaddr >> 12) & 0xfff); gpte 392 arch/powerpc/kvm/book3s_64_mmu_hv.c gpte->may_read = hpte_read_permission(pp, key); gpte 393 arch/powerpc/kvm/book3s_64_mmu_hv.c gpte->may_write = hpte_write_permission(pp, key); gpte 394 arch/powerpc/kvm/book3s_64_mmu_hv.c gpte->may_execute = gpte->may_read && !(gr & (HPTE_R_N | HPTE_R_G)); gpte 400 arch/powerpc/kvm/book3s_64_mmu_hv.c gpte->may_read = 0; gpte 402 arch/powerpc/kvm/book3s_64_mmu_hv.c gpte->may_write = 0; gpte 406 arch/powerpc/kvm/book3s_64_mmu_hv.c gpte->raddr = kvmppc_mmu_get_real_addr(v, gr, eaddr); gpte 128 arch/powerpc/kvm/book3s_64_mmu_radix.c struct kvmppc_pte *gpte, u64 root, gpte 193 arch/powerpc/kvm/book3s_64_mmu_radix.c gpte->page_size = ps; gpte 194 arch/powerpc/kvm/book3s_64_mmu_radix.c gpte->page_shift = offset; gpte 196 arch/powerpc/kvm/book3s_64_mmu_radix.c gpte->eaddr = eaddr; gpte 197 arch/powerpc/kvm/book3s_64_mmu_radix.c gpte->raddr = gpa; gpte 200 arch/powerpc/kvm/book3s_64_mmu_radix.c gpte->may_read = !!(pte & _PAGE_READ); gpte 201 arch/powerpc/kvm/book3s_64_mmu_radix.c gpte->may_write = !!(pte & _PAGE_WRITE); gpte 202 arch/powerpc/kvm/book3s_64_mmu_radix.c gpte->may_execute = !!(pte & _PAGE_EXEC); gpte 204 arch/powerpc/kvm/book3s_64_mmu_radix.c gpte->rc = pte & (_PAGE_ACCESSED | _PAGE_DIRTY); gpte 221 arch/powerpc/kvm/book3s_64_mmu_radix.c struct kvmppc_pte *gpte, u64 table, gpte 246 arch/powerpc/kvm/book3s_64_mmu_radix.c return kvmppc_mmu_walk_radix_tree(vcpu, eaddr, gpte, root, pte_ret_p); gpte 250 arch/powerpc/kvm/book3s_64_mmu_radix.c struct kvmppc_pte *gpte, bool data, bool iswrite) gpte 268 arch/powerpc/kvm/book3s_64_mmu_radix.c ret = kvmppc_mmu_radix_translate_table(vcpu, eaddr, gpte, gpte 276 arch/powerpc/kvm/book3s_64_mmu_radix.c gpte->may_read = 0; gpte 277 arch/powerpc/kvm/book3s_64_mmu_radix.c gpte->may_write = 0; gpte 278 arch/powerpc/kvm/book3s_64_mmu_radix.c gpte->may_execute = 0; gpte 284 arch/powerpc/kvm/book3s_64_mmu_radix.c gpte->may_read = 0; gpte 286 arch/powerpc/kvm/book3s_64_mmu_radix.c gpte->may_write = 0; gpte 288 arch/powerpc/kvm/book3s_64_mmu_radix.c gpte->may_execute = 0; gpte 1198 arch/powerpc/kvm/book3s_hv_nested.c struct kvmppc_pte gpte, gpte 1210 arch/powerpc/kvm/book3s_hv_nested.c if (pgflags & ~gpte.rc) gpte 1216 arch/powerpc/kvm/book3s_hv_nested.c gpte.raddr, kvm->arch.lpid); gpte 1267 arch/powerpc/kvm/book3s_hv_nested.c struct kvmppc_pte gpte; gpte 1290 arch/powerpc/kvm/book3s_hv_nested.c ret = kvmhv_translate_addr_nested(vcpu, gp, n_gpa, dsisr, &gpte); gpte 1306 arch/powerpc/kvm/book3s_hv_nested.c ret = kvmhv_handle_nested_set_rc(vcpu, gp, n_gpa, gpte, dsisr); gpte 1323 arch/powerpc/kvm/book3s_hv_nested.c l1_shift = gpte.page_shift; gpte 1330 arch/powerpc/kvm/book3s_hv_nested.c gpa = gpte.raddr; gpte 1388 arch/powerpc/kvm/book3s_hv_nested.c perm |= gpte.may_read ? 0UL : _PAGE_READ; gpte 1389 arch/powerpc/kvm/book3s_hv_nested.c perm |= gpte.may_write ? 0UL : _PAGE_WRITE; gpte 1390 arch/powerpc/kvm/book3s_hv_nested.c perm |= gpte.may_execute ? 0UL : _PAGE_EXEC; gpte 1392 arch/powerpc/kvm/book3s_hv_nested.c perm |= (gpte.rc & _PAGE_ACCESSED) ? 0UL : _PAGE_ACCESSED; gpte 1393 arch/powerpc/kvm/book3s_hv_nested.c perm |= ((gpte.rc & _PAGE_DIRTY) && writing) ? 0UL : _PAGE_DIRTY; gpte 638 arch/x86/kvm/mmu.c static gfn_t pse36_gfn_delta(u32 gpte) gpte 642 arch/x86/kvm/mmu.c return (gpte & PT32_DIR_PSE36_MASK) << shift; gpte 4016 arch/x86/kvm/mmu.c static bool is_rsvd_bits_set(struct kvm_mmu *mmu, u64 gpte, int level) gpte 4018 arch/x86/kvm/mmu.c return __is_rsvd_bits_set(&mmu->guest_rsvd_check, gpte, level); gpte 4481 arch/x86/kvm/mmu.c unsigned level, unsigned gpte) gpte 4488 arch/x86/kvm/mmu.c gpte &= level - mmu->last_nonleaf_level; gpte 4495 arch/x86/kvm/mmu.c gpte |= level - PT_PAGE_TABLE_LEVEL - 1; gpte 4497 arch/x86/kvm/mmu.c return gpte & PT_PAGE_SIZE_MASK; gpte 99 arch/x86/kvm/paging_tmpl.h static gfn_t gpte_to_gfn_lvl(pt_element_t gpte, int lvl) gpte 101 arch/x86/kvm/paging_tmpl.h return (gpte & PT_LVL_ADDR_MASK(lvl)) >> PAGE_SHIFT; gpte 105 arch/x86/kvm/paging_tmpl.h unsigned gpte) gpte 117 arch/x86/kvm/paging_tmpl.h mask |= (gpte >> (PT_GUEST_DIRTY_SHIFT - PT_WRITABLE_SHIFT)) & gpte 176 arch/x86/kvm/paging_tmpl.h u64 gpte) gpte 178 arch/x86/kvm/paging_tmpl.h if (is_rsvd_bits_set(vcpu->arch.mmu, gpte, PT_PAGE_TABLE_LEVEL)) gpte 181 arch/x86/kvm/paging_tmpl.h if (!FNAME(is_present_gpte)(gpte)) gpte 186 arch/x86/kvm/paging_tmpl.h !(gpte & PT_GUEST_ACCESSED_MASK)) gpte 202 arch/x86/kvm/paging_tmpl.h static inline unsigned FNAME(gpte_access)(u64 gpte) gpte 206 arch/x86/kvm/paging_tmpl.h access = ((gpte & VMX_EPT_WRITABLE_MASK) ? ACC_WRITE_MASK : 0) | gpte 207 arch/x86/kvm/paging_tmpl.h ((gpte & VMX_EPT_EXECUTABLE_MASK) ? ACC_EXEC_MASK : 0) | gpte 208 arch/x86/kvm/paging_tmpl.h ((gpte & VMX_EPT_READABLE_MASK) ? ACC_USER_MASK : 0); gpte 212 arch/x86/kvm/paging_tmpl.h access = gpte & (PT_WRITABLE_MASK | PT_USER_MASK | PT_PRESENT_MASK); gpte 214 arch/x86/kvm/paging_tmpl.h access ^= (gpte >> PT64_NX_SHIFT); gpte 282 arch/x86/kvm/paging_tmpl.h static inline unsigned FNAME(gpte_pkeys)(struct kvm_vcpu *vcpu, u64 gpte) gpte 286 arch/x86/kvm/paging_tmpl.h pte_t pte = {.pte = gpte}; gpte 517 arch/x86/kvm/paging_tmpl.h u64 *spte, pt_element_t gpte, bool no_dirty_log) gpte 523 arch/x86/kvm/paging_tmpl.h if (FNAME(prefetch_invalid_gpte)(vcpu, sp, spte, gpte)) gpte 526 arch/x86/kvm/paging_tmpl.h pgprintk("%s: gpte %llx spte %p\n", __func__, (u64)gpte, spte); gpte 528 arch/x86/kvm/paging_tmpl.h gfn = gpte_to_gfn(gpte); gpte 529 arch/x86/kvm/paging_tmpl.h pte_access = sp->role.access & FNAME(gpte_access)(gpte); gpte 530 arch/x86/kvm/paging_tmpl.h FNAME(protect_clean_gpte)(vcpu->arch.mmu, &pte_access, gpte); gpte 550 arch/x86/kvm/paging_tmpl.h pt_element_t gpte = *(const pt_element_t *)pte; gpte 552 arch/x86/kvm/paging_tmpl.h FNAME(prefetch_gpte)(vcpu, sp, spte, gpte, false); gpte 919 arch/x86/kvm/paging_tmpl.h pt_element_t gpte; gpte 935 arch/x86/kvm/paging_tmpl.h if (kvm_vcpu_read_guest_atomic(vcpu, pte_gpa, &gpte, gpte 939 arch/x86/kvm/paging_tmpl.h FNAME(update_pte)(vcpu, sp, sptep, &gpte); gpte 1021 arch/x86/kvm/paging_tmpl.h pt_element_t gpte; gpte 1030 arch/x86/kvm/paging_tmpl.h if (kvm_vcpu_read_guest_atomic(vcpu, pte_gpa, &gpte, gpte 1034 arch/x86/kvm/paging_tmpl.h if (FNAME(prefetch_invalid_gpte)(vcpu, sp, &sp->spt[i], gpte)) { gpte 1045 arch/x86/kvm/paging_tmpl.h gfn = gpte_to_gfn(gpte); gpte 1047 arch/x86/kvm/paging_tmpl.h pte_access &= FNAME(gpte_access)(gpte); gpte 1048 arch/x86/kvm/paging_tmpl.h FNAME(protect_clean_gpte)(vcpu->arch.mmu, &pte_access, gpte);