gva_t 800 arch/mips/include/asm/kvm_host.h gpa_t (*gva_to_gpa)(gva_t gva); gva_t 23 arch/mips/kvm/trap_emul.c static gpa_t kvm_trap_emul_gva_to_gpa_cb(gva_t gva) gva_t 26 arch/mips/kvm/trap_emul.c gva_t kseg = KSEGX(gva); gva_t 27 arch/mips/kvm/trap_emul.c gva_t gkseg = KVM_GUEST_KSEGX(gva); gva_t 183 arch/mips/kvm/vz.c static gpa_t kvm_vz_gva_to_gpa_cb(gva_t gva) gva_t 161 arch/powerpc/include/asm/kvm_book3s.h extern long kvmppc_hv_find_lock_hpte(struct kvm *kvm, gva_t eaddr, gva_t 164 arch/powerpc/include/asm/kvm_book3s.h unsigned long gpa, gva_t ea, int is_store); gva_t 181 arch/powerpc/include/asm/kvm_book3s.h gva_t eaddr, void *to, void *from, gva_t 183 arch/powerpc/include/asm/kvm_book3s.h extern long kvmhv_copy_from_guest_radix(struct kvm_vcpu *vcpu, gva_t eaddr, gva_t 185 arch/powerpc/include/asm/kvm_book3s.h extern long kvmhv_copy_to_guest_radix(struct kvm_vcpu *vcpu, gva_t eaddr, gva_t 187 arch/powerpc/include/asm/kvm_book3s.h extern int kvmppc_mmu_walk_radix_tree(struct kvm_vcpu *vcpu, gva_t eaddr, gva_t 190 arch/powerpc/include/asm/kvm_book3s.h extern int kvmppc_mmu_radix_translate_table(struct kvm_vcpu *vcpu, gva_t eaddr, gva_t 193 arch/powerpc/include/asm/kvm_book3s.h extern int kvmppc_mmu_radix_xlate(struct kvm_vcpu *vcpu, gva_t eaddr, gva_t 396 arch/powerpc/include/asm/kvm_host.h int (*slbfee)(struct kvm_vcpu *vcpu, gva_t eaddr, ulong *ret_slb); gva_t 402 arch/powerpc/include/asm/kvm_host.h int (*xlate)(struct kvm_vcpu *vcpu, gva_t eaddr, gva_t 407 arch/powerpc/include/asm/kvm_host.h u64 (*ea_to_vp)(struct kvm_vcpu *vcpu, gva_t eaddr, bool data); gva_t 704 arch/powerpc/include/asm/kvm_host.h gva_t vaddr_accessed; gva_t 112 arch/powerpc/include/asm/kvm_ppc.h extern int kvmppc_mmu_dtlb_index(struct kvm_vcpu *vcpu, gva_t eaddr); gva_t 113 arch/powerpc/include/asm/kvm_ppc.h extern int kvmppc_mmu_itlb_index(struct kvm_vcpu *vcpu, gva_t eaddr); gva_t 115 arch/powerpc/include/asm/kvm_ppc.h gva_t eaddr); gva_t 69 arch/powerpc/kvm/book3s_32_mmu.c static int kvmppc_mmu_book3s_32_xlate_bat(struct kvm_vcpu *vcpu, gva_t eaddr, gva_t 75 arch/powerpc/kvm/book3s_32_mmu.c static u32 find_sr(struct kvm_vcpu *vcpu, gva_t eaddr) gva_t 80 arch/powerpc/kvm/book3s_32_mmu.c static u64 kvmppc_mmu_book3s_32_ea_to_vp(struct kvm_vcpu *vcpu, gva_t eaddr, gva_t 99 arch/powerpc/kvm/book3s_32_mmu.c u32 sre, gva_t eaddr, gva_t 126 arch/powerpc/kvm/book3s_32_mmu.c static u32 kvmppc_mmu_book3s_32_get_ptem(u32 sre, gva_t eaddr, bool primary) gva_t 132 arch/powerpc/kvm/book3s_32_mmu.c static int kvmppc_mmu_book3s_32_xlate_bat(struct kvm_vcpu *vcpu, gva_t eaddr, gva_t 187 arch/powerpc/kvm/book3s_32_mmu.c static int kvmppc_mmu_book3s_32_xlate_pte(struct kvm_vcpu *vcpu, gva_t eaddr, gva_t 295 arch/powerpc/kvm/book3s_32_mmu.c static int kvmppc_mmu_book3s_32_xlate(struct kvm_vcpu *vcpu, gva_t eaddr, gva_t 43 arch/powerpc/kvm/book3s_64_mmu.c gva_t eaddr) gva_t 87 arch/powerpc/kvm/book3s_64_mmu.c static u64 kvmppc_slb_calc_vpn(struct kvmppc_slb *slb, gva_t eaddr) gva_t 95 arch/powerpc/kvm/book3s_64_mmu.c static u64 kvmppc_mmu_book3s_64_ea_to_vp(struct kvm_vcpu *vcpu, gva_t eaddr, gva_t 123 arch/powerpc/kvm/book3s_64_mmu.c static u32 kvmppc_mmu_book3s_64_get_page(struct kvmppc_slb *slbe, gva_t eaddr) gva_t 131 arch/powerpc/kvm/book3s_64_mmu.c struct kvmppc_slb *slbe, gva_t eaddr, gva_t 169 arch/powerpc/kvm/book3s_64_mmu.c static u64 kvmppc_mmu_book3s_64_get_avpn(struct kvmppc_slb *slbe, gva_t eaddr) gva_t 205 arch/powerpc/kvm/book3s_64_mmu.c static int kvmppc_mmu_book3s_64_xlate(struct kvm_vcpu *vcpu, gva_t eaddr, gva_t 427 arch/powerpc/kvm/book3s_64_mmu.c static int kvmppc_mmu_book3s_64_slbfee(struct kvm_vcpu *vcpu, gva_t eaddr, gva_t 311 arch/powerpc/kvm/book3s_64_mmu_hv.c gva_t eaddr) gva_t 340 arch/powerpc/kvm/book3s_64_mmu_hv.c static int kvmppc_mmu_book3s_64_hv_xlate(struct kvm_vcpu *vcpu, gva_t eaddr, gva_t 429 arch/powerpc/kvm/book3s_64_mmu_hv.c unsigned long gpa, gva_t ea, int is_store) gva_t 31 arch/powerpc/kvm/book3s_64_mmu_radix.c gva_t eaddr, void *to, void *from, gva_t 84 arch/powerpc/kvm/book3s_64_mmu_radix.c static long kvmhv_copy_tofrom_guest_radix(struct kvm_vcpu *vcpu, gva_t eaddr, gva_t 107 arch/powerpc/kvm/book3s_64_mmu_radix.c long kvmhv_copy_from_guest_radix(struct kvm_vcpu *vcpu, gva_t eaddr, void *to, gva_t 120 arch/powerpc/kvm/book3s_64_mmu_radix.c long kvmhv_copy_to_guest_radix(struct kvm_vcpu *vcpu, gva_t eaddr, void *from, gva_t 127 arch/powerpc/kvm/book3s_64_mmu_radix.c int kvmppc_mmu_walk_radix_tree(struct kvm_vcpu *vcpu, gva_t eaddr, gva_t 220 arch/powerpc/kvm/book3s_64_mmu_radix.c int kvmppc_mmu_radix_translate_table(struct kvm_vcpu *vcpu, gva_t eaddr, gva_t 249 arch/powerpc/kvm/book3s_64_mmu_radix.c int kvmppc_mmu_radix_xlate(struct kvm_vcpu *vcpu, gva_t eaddr, gva_t 478 arch/powerpc/kvm/book3s_hv_nested.c gva_t eaddr = kvmppc_get_gpr(vcpu, 6); gva_t 1115 arch/powerpc/kvm/book3s_hv_rm_mmu.c long kvmppc_hv_find_lock_hpte(struct kvm *kvm, gva_t eaddr, unsigned long slb_v, gva_t 129 arch/powerpc/kvm/e500.h int kvmppc_e500_emul_tlbivax(struct kvm_vcpu *vcpu, gva_t ea); gva_t 130 arch/powerpc/kvm/e500.h int kvmppc_e500_emul_tlbilx(struct kvm_vcpu *vcpu, int type, gva_t ea); gva_t 131 arch/powerpc/kvm/e500.h int kvmppc_e500_emul_tlbsx(struct kvm_vcpu *vcpu, gva_t ea); gva_t 156 arch/powerpc/kvm/e500.h static inline gva_t get_tlb_eaddr(const struct kvm_book3e_206_tlb_entry *tlbe) gva_t 167 arch/powerpc/kvm/e500.h static inline gva_t get_tlb_end(const struct kvm_book3e_206_tlb_entry *tlbe) gva_t 135 arch/powerpc/kvm/e500_emulate.c gva_t ea; gva_t 48 arch/powerpc/kvm/e500_mmu.c static int tlb0_set_base(gva_t addr, int sets, int ways) gva_t 58 arch/powerpc/kvm/e500_mmu.c static int gtlb0_set_base(struct kvmppc_vcpu_e500 *vcpu_e500, gva_t addr) gva_t 81 arch/powerpc/kvm/e500_mmu.c gva_t eaddr, int tlbsel, unsigned int pid, int as) gva_t 127 arch/powerpc/kvm/e500_mmu.c gva_t eaddr, int as) gva_t 155 arch/powerpc/kvm/e500_mmu.c gva_t eaddr; gva_t 245 arch/powerpc/kvm/e500_mmu.c int kvmppc_e500_emul_tlbivax(struct kvm_vcpu *vcpu, gva_t ea) gva_t 293 arch/powerpc/kvm/e500_mmu.c gva_t ea) gva_t 307 arch/powerpc/kvm/e500_mmu.c int kvmppc_e500_emul_tlbilx(struct kvm_vcpu *vcpu, int type, gva_t ea) gva_t 341 arch/powerpc/kvm/e500_mmu.c int kvmppc_e500_emul_tlbsx(struct kvm_vcpu *vcpu, gva_t ea) gva_t 454 arch/powerpc/kvm/e500_mmu.c gva_t eaddr, unsigned int pid, int as) gva_t 473 arch/powerpc/kvm/e500_mmu.c gva_t eaddr; gva_t 495 arch/powerpc/kvm/e500_mmu.c int kvmppc_mmu_itlb_index(struct kvm_vcpu *vcpu, gva_t eaddr) gva_t 502 arch/powerpc/kvm/e500_mmu.c int kvmppc_mmu_dtlb_index(struct kvm_vcpu *vcpu, gva_t eaddr) gva_t 524 arch/powerpc/kvm/e500_mmu.c gva_t eaddr) gva_t 628 arch/powerpc/kvm/e500_mmu_host.c gva_t geaddr; gva_t 59 arch/powerpc/kvm/e500mc.c gva_t eaddr; gva_t 393 arch/x86/include/asm/kvm_host.h void (*invlpg)(struct kvm_vcpu *vcpu, gva_t gva, hpa_t root_hpa); gva_t 1074 arch/x86/include/asm/kvm_host.h void (*tlb_flush_gva)(struct kvm_vcpu *vcpu, gva_t addr); gva_t 1433 arch/x86/include/asm/kvm_host.h int kvm_mmu_unprotect_page_virt(struct kvm_vcpu *vcpu, gva_t gva); gva_t 1442 arch/x86/include/asm/kvm_host.h gpa_t kvm_mmu_gva_to_gpa_read(struct kvm_vcpu *vcpu, gva_t gva, gva_t 1444 arch/x86/include/asm/kvm_host.h gpa_t kvm_mmu_gva_to_gpa_fetch(struct kvm_vcpu *vcpu, gva_t gva, gva_t 1446 arch/x86/include/asm/kvm_host.h gpa_t kvm_mmu_gva_to_gpa_write(struct kvm_vcpu *vcpu, gva_t gva, gva_t 1448 arch/x86/include/asm/kvm_host.h gpa_t kvm_mmu_gva_to_gpa_system(struct kvm_vcpu *vcpu, gva_t gva, gva_t 1457 arch/x86/include/asm/kvm_host.h void kvm_mmu_invlpg(struct kvm_vcpu *vcpu, gva_t gva); gva_t 1458 arch/x86/include/asm/kvm_host.h void kvm_mmu_invpcid_gva(struct kvm_vcpu *vcpu, gva_t gva, unsigned long pcid); gva_t 2081 arch/x86/kvm/emulate.c gva_t cs_addr; gva_t 2082 arch/x86/kvm/emulate.c gva_t eip_addr; gva_t 2230 arch/x86/kvm/mmu.c static void nonpaging_invlpg(struct kvm_vcpu *vcpu, gva_t gva, hpa_t root) gva_t 2552 arch/x86/kvm/mmu.c gva_t gaddr, gva_t 3431 arch/x86/kvm/mmu.c static bool handle_abnormal_pfn(struct kvm_vcpu *vcpu, gva_t gva, gfn_t gfn, gva_t 4141 arch/x86/kvm/mmu.c static void shadow_page_table_clear_flood(struct kvm_vcpu *vcpu, gva_t addr) gva_t 5499 arch/x86/kvm/mmu.c int kvm_mmu_unprotect_page_virt(struct kvm_vcpu *vcpu, gva_t gva) gva_t 5610 arch/x86/kvm/mmu.c void kvm_mmu_invlpg(struct kvm_vcpu *vcpu, gva_t gva) gva_t 5641 arch/x86/kvm/mmu.c void kvm_mmu_invpcid_gva(struct kvm_vcpu *vcpu, gva_t gva, unsigned long pcid) gva_t 507 arch/x86/kvm/paging_tmpl.h struct kvm_vcpu *vcpu, gva_t addr, gva_t 892 arch/x86/kvm/paging_tmpl.h static void FNAME(invlpg)(struct kvm_vcpu *vcpu, gva_t gva, hpa_t root_hpa) gva_t 5537 arch/x86/kvm/svm.c static void svm_flush_tlb_gva(struct kvm_vcpu *vcpu, gva_t gva) gva_t 772 arch/x86/kvm/trace.h TP_PROTO(gva_t gva, gpa_t gpa, bool write, bool gpa_match), gva_t 776 arch/x86/kvm/trace.h __field(gva_t, gva) gva_t 4208 arch/x86/kvm/vmx/nested.c u32 vmx_instruction_info, bool wr, int len, gva_t *ret) gva_t 4210 arch/x86/kvm/vmx/nested.c gva_t off; gva_t 4240 arch/x86/kvm/vmx/nested.c off = (gva_t)sign_extend64(off, 31); gva_t 4242 arch/x86/kvm/vmx/nested.c off = (gva_t)sign_extend64(off, 15); gva_t 4333 arch/x86/kvm/vmx/nested.c gva_t gva; gva_t 4607 arch/x86/kvm/vmx/nested.c gva_t gva = 0; gva_t 4690 arch/x86/kvm/vmx/nested.c gva_t gva; gva_t 4880 arch/x86/kvm/vmx/nested.c gva_t gva; gva_t 4906 arch/x86/kvm/vmx/nested.c gva_t gva; gva_t 4963 arch/x86/kvm/vmx/nested.c gva_t gva; gva_t 34 arch/x86/kvm/vmx/nested.h u32 vmx_instruction_info, bool wr, int len, gva_t *ret); gva_t 21 arch/x86/kvm/vmx/ops.h void invvpid_error(unsigned long ext, u16 vpid, gva_t gva); gva_t 251 arch/x86/kvm/vmx/ops.h static inline void __invvpid(unsigned long ext, u16 vpid, gva_t gva) gva_t 271 arch/x86/kvm/vmx/ops.h static inline bool vpid_sync_vcpu_addr(int vpid, gva_t addr) gva_t 381 arch/x86/kvm/vmx/vmx.c noinline void invvpid_error(unsigned long ext, u16 vpid, gva_t gva) gva_t 2812 arch/x86/kvm/vmx/vmx.c static void vmx_flush_tlb_gva(struct kvm_vcpu *vcpu, gva_t addr) gva_t 5389 arch/x86/kvm/vmx/vmx.c gva_t gva; gva_t 5356 arch/x86/kvm/x86.c gpa_t kvm_mmu_gva_to_gpa_read(struct kvm_vcpu *vcpu, gva_t gva, gva_t 5363 arch/x86/kvm/x86.c gpa_t kvm_mmu_gva_to_gpa_fetch(struct kvm_vcpu *vcpu, gva_t gva, gva_t 5371 arch/x86/kvm/x86.c gpa_t kvm_mmu_gva_to_gpa_write(struct kvm_vcpu *vcpu, gva_t gva, gva_t 5380 arch/x86/kvm/x86.c gpa_t kvm_mmu_gva_to_gpa_system(struct kvm_vcpu *vcpu, gva_t gva, gva_t 5386 arch/x86/kvm/x86.c static int kvm_read_guest_virt_helper(gva_t addr, void *val, unsigned int bytes, gva_t 5419 arch/x86/kvm/x86.c gva_t addr, void *val, unsigned int bytes, gva_t 5445 arch/x86/kvm/x86.c gva_t addr, void *val, unsigned int bytes, gva_t 5463 arch/x86/kvm/x86.c gva_t addr, void *val, unsigned int bytes, gva_t 5484 arch/x86/kvm/x86.c static int kvm_write_guest_virt_helper(gva_t addr, void *val, unsigned int bytes, gva_t 5515 arch/x86/kvm/x86.c static int emulator_write_std(struct x86_emulate_ctxt *ctxt, gva_t addr, void *val, gva_t 5529 arch/x86/kvm/x86.c int kvm_write_guest_virt_system(struct kvm_vcpu *vcpu, gva_t addr, void *val, gva_t 187 arch/x86/kvm/x86.h gva_t gva, gfn_t gfn, unsigned access) gva_t 213 arch/x86/kvm/x86.h #define MMIO_GVA_ANY (~(gva_t)0) gva_t 215 arch/x86/kvm/x86.h static inline void vcpu_clear_mmio_info(struct kvm_vcpu *vcpu, gva_t gva) gva_t 270 arch/x86/kvm/x86.h gva_t addr, void *val, unsigned int bytes, gva_t 274 arch/x86/kvm/x86.h gva_t addr, void *val, unsigned int bytes,