Lines Matching refs:gpa

1968 	gpa_t gpa = data & ~0x3f;  in kvm_pv_enable_async_pf()  local
1982 if (kvm_gfn_to_hva_cache_init(vcpu->kvm, &vcpu->arch.apf.data, gpa, in kvm_pv_enable_async_pf()
4107 gpa_t translate_nested_gpa(struct kvm_vcpu *vcpu, gpa_t gpa, u32 access, in translate_nested_gpa() argument
4116 t_gpa = vcpu->arch.mmu.gva_to_gpa(vcpu, gpa, access, exception); in translate_nested_gpa()
4159 gpa_t gpa = vcpu->arch.walk_mmu->gva_to_gpa(vcpu, addr, access, in kvm_read_guest_virt_helper() local
4165 if (gpa == UNMAPPED_GVA) in kvm_read_guest_virt_helper()
4167 ret = kvm_vcpu_read_guest_page(vcpu, gpa >> PAGE_SHIFT, data, in kvm_read_guest_virt_helper()
4193 gpa_t gpa = vcpu->arch.walk_mmu->gva_to_gpa(vcpu, addr, access|PFERR_FETCH_MASK, in kvm_fetch_guest_virt() local
4195 if (unlikely(gpa == UNMAPPED_GVA)) in kvm_fetch_guest_virt()
4201 ret = kvm_vcpu_read_guest_page(vcpu, gpa >> PAGE_SHIFT, val, in kvm_fetch_guest_virt()
4248 gpa_t gpa = vcpu->arch.walk_mmu->gva_to_gpa(vcpu, addr, in kvm_write_guest_virt_system() local
4255 if (gpa == UNMAPPED_GVA) in kvm_write_guest_virt_system()
4257 ret = kvm_vcpu_write_guest(vcpu, gpa, data, towrite); in kvm_write_guest_virt_system()
4273 gpa_t *gpa, struct x86_exception *exception, in vcpu_mmio_gva_to_gpa() argument
4282 *gpa = vcpu->arch.mmio_gfn << PAGE_SHIFT | in vcpu_mmio_gva_to_gpa()
4284 trace_vcpu_match_mmio(gva, *gpa, write, false); in vcpu_mmio_gva_to_gpa()
4288 *gpa = vcpu->arch.walk_mmu->gva_to_gpa(vcpu, gva, access, exception); in vcpu_mmio_gva_to_gpa()
4290 if (*gpa == UNMAPPED_GVA) in vcpu_mmio_gva_to_gpa()
4294 if ((*gpa & PAGE_MASK) == APIC_DEFAULT_PHYS_BASE) in vcpu_mmio_gva_to_gpa()
4297 if (vcpu_match_mmio_gpa(vcpu, *gpa)) { in vcpu_mmio_gva_to_gpa()
4298 trace_vcpu_match_mmio(gva, *gpa, write, true); in vcpu_mmio_gva_to_gpa()
4305 int emulator_write_phys(struct kvm_vcpu *vcpu, gpa_t gpa, in emulator_write_phys() argument
4310 ret = kvm_vcpu_write_guest(vcpu, gpa, val, bytes); in emulator_write_phys()
4313 kvm_mmu_pte_write(vcpu, gpa, val, bytes); in emulator_write_phys()
4320 int (*read_write_emulate)(struct kvm_vcpu *vcpu, gpa_t gpa,
4322 int (*read_write_mmio)(struct kvm_vcpu *vcpu, gpa_t gpa,
4324 int (*read_write_exit_mmio)(struct kvm_vcpu *vcpu, gpa_t gpa,
4333 vcpu->mmio_fragments[0].gpa, *(u64 *)val); in read_prepare()
4341 static int read_emulate(struct kvm_vcpu *vcpu, gpa_t gpa, in read_emulate() argument
4344 return !kvm_vcpu_read_guest(vcpu, gpa, val, bytes); in read_emulate()
4347 static int write_emulate(struct kvm_vcpu *vcpu, gpa_t gpa, in write_emulate() argument
4350 return emulator_write_phys(vcpu, gpa, val, bytes); in write_emulate()
4353 static int write_mmio(struct kvm_vcpu *vcpu, gpa_t gpa, int bytes, void *val) in write_mmio() argument
4355 trace_kvm_mmio(KVM_TRACE_MMIO_WRITE, bytes, gpa, *(u64 *)val); in write_mmio()
4356 return vcpu_mmio_write(vcpu, gpa, bytes, val); in write_mmio()
4359 static int read_exit_mmio(struct kvm_vcpu *vcpu, gpa_t gpa, in read_exit_mmio() argument
4362 trace_kvm_mmio(KVM_TRACE_MMIO_READ_UNSATISFIED, bytes, gpa, 0); in read_exit_mmio()
4366 static int write_exit_mmio(struct kvm_vcpu *vcpu, gpa_t gpa, in write_exit_mmio() argument
4395 gpa_t gpa; in emulator_read_write_onepage() local
4400 ret = vcpu_mmio_gva_to_gpa(vcpu, addr, &gpa, exception, write); in emulator_read_write_onepage()
4409 if (ops->read_write_emulate(vcpu, gpa, val, bytes)) in emulator_read_write_onepage()
4416 handled = ops->read_write_mmio(vcpu, gpa, bytes, val); in emulator_read_write_onepage()
4420 gpa += handled; in emulator_read_write_onepage()
4426 frag->gpa = gpa; in emulator_read_write_onepage()
4439 gpa_t gpa; in emulator_read_write() local
4473 gpa = vcpu->mmio_fragments[0].gpa; in emulator_read_write()
4481 vcpu->run->mmio.phys_addr = gpa; in emulator_read_write()
4483 return ops->read_write_exit_mmio(vcpu, gpa, val, bytes); in emulator_read_write()
4524 gpa_t gpa; in emulator_cmpxchg_emulated() local
4533 gpa = kvm_mmu_gva_to_gpa_write(vcpu, addr, NULL); in emulator_cmpxchg_emulated()
4535 if (gpa == UNMAPPED_GVA || in emulator_cmpxchg_emulated()
4536 (gpa & PAGE_MASK) == APIC_DEFAULT_PHYS_BASE) in emulator_cmpxchg_emulated()
4539 if (((gpa + bytes - 1) & PAGE_MASK) != (gpa & PAGE_MASK)) in emulator_cmpxchg_emulated()
4542 page = kvm_vcpu_gfn_to_page(vcpu, gpa >> PAGE_SHIFT); in emulator_cmpxchg_emulated()
4547 kaddr += offset_in_page(gpa); in emulator_cmpxchg_emulated()
4570 kvm_vcpu_mark_page_dirty(vcpu, gpa >> PAGE_SHIFT); in emulator_cmpxchg_emulated()
4571 kvm_mmu_pte_write(vcpu, gpa, new, bytes); in emulator_cmpxchg_emulated()
5114 gpa_t gpa = cr2; in reexecute_instruction() local
5125 gpa = kvm_mmu_gva_to_gpa_write(vcpu, cr2, NULL); in reexecute_instruction()
5131 if (gpa == UNMAPPED_GVA) in reexecute_instruction()
5141 pfn = gfn_to_pfn(vcpu->kvm, gpa_to_gfn(gpa)); in reexecute_instruction()
5161 kvm_mmu_unprotect_page(vcpu->kvm, gpa_to_gfn(gpa)); in reexecute_instruction()
5171 kvm_mmu_unprotect_page(vcpu->kvm, gpa_to_gfn(gpa)); in reexecute_instruction()
5185 unsigned long last_retry_eip, last_retry_addr, gpa = cr2; in retry_instruction() local
5218 gpa = kvm_mmu_gva_to_gpa_write(vcpu, cr2, NULL); in retry_instruction()
5220 kvm_mmu_unprotect_page(vcpu->kvm, gpa_to_gfn(gpa)); in retry_instruction()
6766 frag->gpa += len; in complete_emulated_mmio()
6781 run->mmio.phys_addr = frag->gpa; in complete_emulated_mmio()
7151 gpa_t gpa; in kvm_arch_vcpu_ioctl_translate() local
7155 gpa = kvm_mmu_gva_to_gpa_system(vcpu, vaddr, NULL); in kvm_arch_vcpu_ioctl_translate()
7157 tr->physical_address = gpa; in kvm_arch_vcpu_ioctl_translate()
7158 tr->valid = gpa != UNMAPPED_GVA; in kvm_arch_vcpu_ioctl_translate()
7696 int __x86_set_memory_region(struct kvm *kvm, int id, gpa_t gpa, u32 size) in __x86_set_memory_region() argument
7733 m.guest_phys_addr = gpa; in __x86_set_memory_region()
7750 int x86_set_memory_region(struct kvm *kvm, int id, gpa_t gpa, u32 size) in x86_set_memory_region() argument
7755 r = __x86_set_memory_region(kvm, id, gpa, size); in x86_set_memory_region()