Searched refs:gpa (Results 1 - 47 of 47) sorted by relevance

/linux-4.4.14/include/linux/
H A Dkvm_types.h41 * gpa - guest physical address
60 gpa_t gpa; member in struct:gfn_to_hva_cache
H A Dkvm_host.h220 gpa_t gpa; member in struct:kvm_mmio_fragment
621 int kvm_read_guest_atomic(struct kvm *kvm, gpa_t gpa, void *data,
623 int kvm_read_guest(struct kvm *kvm, gpa_t gpa, void *data, unsigned long len);
628 int kvm_write_guest(struct kvm *kvm, gpa_t gpa, const void *data,
633 gpa_t gpa, unsigned long len);
635 int kvm_clear_guest(struct kvm *kvm, gpa_t gpa, unsigned long len);
650 int kvm_vcpu_read_guest_atomic(struct kvm_vcpu *vcpu, gpa_t gpa, void *data,
652 int kvm_vcpu_read_guest(struct kvm_vcpu *vcpu, gpa_t gpa, void *data,
656 int kvm_vcpu_write_guest(struct kvm_vcpu *vcpu, gpa_t gpa, const void *data,
976 static inline gfn_t gpa_to_gfn(gpa_t gpa) gpa_to_gfn() argument
978 return (gfn_t)(gpa >> PAGE_SHIFT); gpa_to_gfn()
986 static inline bool kvm_is_error_gpa(struct kvm *kvm, gpa_t gpa) kvm_is_error_gpa() argument
988 unsigned long hva = gfn_to_hva(kvm, gpa_to_gfn(gpa)); kvm_is_error_gpa()
/linux-4.4.14/arch/s390/kvm/
H A Dgaccess.h127 unsigned long gpa = gra + kvm_s390_get_prefix(vcpu); write_guest_lc() local
129 return kvm_write_guest(vcpu->kvm, gpa, data, len); write_guest_lc()
153 unsigned long gpa = gra + kvm_s390_get_prefix(vcpu); read_guest_lc() local
155 return kvm_read_guest(vcpu->kvm, gpa, data, len); read_guest_lc()
159 ar_t ar, unsigned long *gpa, int write);
244 * @gpa: guest physical (absolute) address
248 * Copy @len bytes from @data (kernel space) to @gpa (guest absolute address).
258 int write_guest_abs(struct kvm_vcpu *vcpu, unsigned long gpa, void *data, write_guest_abs() argument
261 return kvm_write_guest(vcpu->kvm, gpa, data, len); write_guest_abs()
267 * @gpa: guest physical (absolute) address
271 * Copy @len bytes from @gpa (guest absolute address) to @data (kernel space).
281 int read_guest_abs(struct kvm_vcpu *vcpu, unsigned long gpa, void *data, read_guest_abs() argument
284 return kvm_read_guest(vcpu->kvm, gpa, data, len); read_guest_abs()
H A Dgaccess.c514 static int deref_table(struct kvm *kvm, unsigned long gpa, unsigned long *val) deref_table() argument
516 return kvm_read_guest(kvm, gpa, val, sizeof(*val)); deref_table()
523 * @gpa: points to where guest physical (absolute) address should be stored
530 * an addressing exception is indicated and @gpa will not be changed.
532 * Returns: - zero on success; @gpa contains the resulting absolute address
540 unsigned long *gpa, const union asce asce, guest_translate()
690 *gpa = raddr.addr; guest_translate()
758 unsigned long _len, nr_pages, gpa, idx; access_guest() local
781 gpa = *(pages + idx) + (ga & ~PAGE_MASK); access_guest()
782 _len = min(PAGE_SIZE - (gpa & ~PAGE_MASK), len); access_guest()
784 rc = kvm_write_guest(vcpu->kvm, gpa, data, _len); access_guest()
786 rc = kvm_read_guest(vcpu->kvm, gpa, data, _len); access_guest()
801 unsigned long _len, gpa; access_guest_real() local
805 gpa = kvm_s390_real_to_abs(vcpu, gra); access_guest_real()
806 _len = min(PAGE_SIZE - (gpa & ~PAGE_MASK), len); access_guest_real()
808 rc = write_guest_abs(vcpu, gpa, data, _len); access_guest_real()
810 rc = read_guest_abs(vcpu, gpa, data, _len); access_guest_real()
828 unsigned long *gpa, int write) guest_translate_address()
850 rc = guest_translate(vcpu, gva, gpa, asce, write); guest_translate_address()
858 *gpa = kvm_s390_real_to_abs(vcpu, gva); guest_translate_address()
859 if (kvm_is_error_gpa(vcpu->kvm, *gpa)) guest_translate_address()
872 unsigned long gpa; check_gva_range() local
879 rc = guest_translate_address(vcpu, gva, ar, &gpa, is_write); check_gva_range()
539 guest_translate(struct kvm_vcpu *vcpu, unsigned long gva, unsigned long *gpa, const union asce asce, int write) guest_translate() argument
827 guest_translate_address(struct kvm_vcpu *vcpu, unsigned long gva, ar_t ar, unsigned long *gpa, int write) guest_translate_address() argument
H A Dkvm-s390.c1881 * @gpa: Guest physical address
1888 long kvm_arch_fault_in_page(struct kvm_vcpu *vcpu, gpa_t gpa, int writable) kvm_arch_fault_in_page() argument
1890 return gmap_fault(vcpu->arch.gmap, gpa, kvm_arch_fault_in_page()
2062 gpa_t gpa = current->thread.gmap_addr; vcpu_post_run() local
2063 rc = kvm_arch_fault_in_page(vcpu, gpa, 1); vcpu_post_run()
2230 int kvm_s390_store_status_unloaded(struct kvm_vcpu *vcpu, unsigned long gpa) kvm_s390_store_status_unloaded() argument
2239 if (gpa == KVM_S390_STORE_STATUS_NOADDR) { kvm_s390_store_status_unloaded()
2242 gpa = 0; kvm_s390_store_status_unloaded()
2243 } else if (gpa == KVM_S390_STORE_STATUS_PREFIXED) { kvm_s390_store_status_unloaded()
2246 gpa = px; kvm_s390_store_status_unloaded()
2248 gpa -= __LC_FPREGS_SAVE_AREA; kvm_s390_store_status_unloaded()
2253 rc = write_guest_abs(vcpu, gpa + __LC_FPREGS_SAVE_AREA, kvm_s390_store_status_unloaded()
2256 rc = write_guest_abs(vcpu, gpa + __LC_FPREGS_SAVE_AREA, kvm_s390_store_status_unloaded()
2259 rc |= write_guest_abs(vcpu, gpa + __LC_GPREGS_SAVE_AREA, kvm_s390_store_status_unloaded()
2261 rc |= write_guest_abs(vcpu, gpa + __LC_PSW_SAVE_AREA, kvm_s390_store_status_unloaded()
2263 rc |= write_guest_abs(vcpu, gpa + __LC_PREFIX_SAVE_AREA, kvm_s390_store_status_unloaded()
2265 rc |= write_guest_abs(vcpu, gpa + __LC_FP_CREG_SAVE_AREA, kvm_s390_store_status_unloaded()
2267 rc |= write_guest_abs(vcpu, gpa + __LC_TOD_PROGREG_SAVE_AREA, kvm_s390_store_status_unloaded()
2269 rc |= write_guest_abs(vcpu, gpa + __LC_CPU_TIMER_SAVE_AREA, kvm_s390_store_status_unloaded()
2272 rc |= write_guest_abs(vcpu, gpa + __LC_CLOCK_COMP_SAVE_AREA, kvm_s390_store_status_unloaded()
2274 rc |= write_guest_abs(vcpu, gpa + __LC_AREGS_SAVE_AREA, kvm_s390_store_status_unloaded()
2276 rc |= write_guest_abs(vcpu, gpa + __LC_CREGS_SAVE_AREA, kvm_s390_store_status_unloaded()
2299 unsigned long gpa) kvm_s390_store_adtl_status_unloaded()
2302 if (!(gpa & ~0x3ff)) kvm_s390_store_adtl_status_unloaded()
2305 return write_guest_abs(vcpu, gpa & ~0x3ff, kvm_s390_store_adtl_status_unloaded()
2298 kvm_s390_store_adtl_status_unloaded(struct kvm_vcpu *vcpu, unsigned long gpa) kvm_s390_store_adtl_status_unloaded() argument
H A Dpriv.c965 unsigned long hva, gpa; handle_tprot() local
984 ret = guest_translate_address(vcpu, address1, ar, &gpa, 1); handle_tprot()
988 ret = guest_translate_address(vcpu, address1, ar, &gpa, 0); handle_tprot()
1001 hva = gfn_to_hva_prot(vcpu->kvm, gpa_to_gfn(gpa), &writable); handle_tprot()
H A Dkvm-s390.h235 long kvm_arch_fault_in_page(struct kvm_vcpu *vcpu, gpa_t gpa, int writable);
/linux-4.4.14/include/trace/events/
H A Dkvm.h207 TP_PROTO(int type, int len, u64 gpa, u64 val),
208 TP_ARGS(type, len, gpa, val),
213 __field( u64, gpa )
220 __entry->gpa = gpa;
224 TP_printk("mmio %s len %u gpa 0x%llx val 0x%llx",
226 __entry->len, __entry->gpa, __entry->val)
/linux-4.4.14/arch/x86/include/asm/uv/
H A Duv_hub.h330 uv_gpa_in_mmr_space(unsigned long gpa) uv_gpa_in_mmr_space() argument
332 return (gpa >> 62) == 0x3UL; uv_gpa_in_mmr_space()
336 static inline unsigned long uv_gpa_to_soc_phys_ram(unsigned long gpa) uv_gpa_to_soc_phys_ram() argument
342 gpa = ((gpa << uv_hub_info->m_shift) >> uv_hub_info->m_shift) | uv_gpa_to_soc_phys_ram()
343 ((gpa >> uv_hub_info->n_lshift) << uv_hub_info->m_val); uv_gpa_to_soc_phys_ram()
344 paddr = gpa & uv_hub_info->gpa_mask; uv_gpa_to_soc_phys_ram()
351 /* gpa -> pnode */ uv_gpa_to_gnode()
352 static inline unsigned long uv_gpa_to_gnode(unsigned long gpa) uv_gpa_to_gnode() argument
354 return gpa >> uv_hub_info->n_lshift; uv_gpa_to_gnode()
357 /* gpa -> pnode */ uv_gpa_to_pnode()
358 static inline int uv_gpa_to_pnode(unsigned long gpa) uv_gpa_to_pnode() argument
362 return uv_gpa_to_gnode(gpa) & n_mask; uv_gpa_to_pnode()
365 /* gpa -> node offset*/ uv_gpa_to_offset()
366 static inline unsigned long uv_gpa_to_offset(unsigned long gpa) uv_gpa_to_offset() argument
368 return (gpa << uv_hub_info->m_shift) >> uv_hub_info->m_shift; uv_gpa_to_offset()
/linux-4.4.14/arch/x86/kvm/
H A Dtrace.h135 TP_PROTO(u64 gpa),
136 TP_ARGS(gpa),
139 __field(u64, gpa)
143 __entry->gpa = gpa;
146 TP_printk("fast mmio at gpa 0x%llx", __entry->gpa)
765 TP_PROTO(gva_t gva, gpa_t gpa, bool write, bool gpa_match),
766 TP_ARGS(gva, gpa, write, gpa_match),
770 __field(gpa_t, gpa)
777 __entry->gpa = gpa;
782 TP_printk("gva %#lx gpa %#llx %s %s", __entry->gva, __entry->gpa,
H A Dmmutrace.h103 __field(__u64, gpa)
107 __entry->gpa = ((u64)table_gfn << PAGE_SHIFT)
111 TP_printk("gpa %llx", __entry->gpa)
H A Dx86.h124 static inline bool vcpu_match_mmio_gpa(struct kvm_vcpu *vcpu, gpa_t gpa) vcpu_match_mmio_gpa() argument
127 vcpu->arch.mmio_gfn == gpa >> PAGE_SHIFT) vcpu_match_mmio_gpa()
H A Dx86.c1968 gpa_t gpa = data & ~0x3f; kvm_pv_enable_async_pf() local
1982 if (kvm_gfn_to_hva_cache_init(vcpu->kvm, &vcpu->arch.apf.data, gpa, kvm_pv_enable_async_pf()
4107 gpa_t translate_nested_gpa(struct kvm_vcpu *vcpu, gpa_t gpa, u32 access, translate_nested_gpa() argument
4116 t_gpa = vcpu->arch.mmu.gva_to_gpa(vcpu, gpa, access, exception); translate_nested_gpa()
4159 gpa_t gpa = vcpu->arch.walk_mmu->gva_to_gpa(vcpu, addr, access, kvm_read_guest_virt_helper() local
4165 if (gpa == UNMAPPED_GVA) kvm_read_guest_virt_helper()
4167 ret = kvm_vcpu_read_guest_page(vcpu, gpa >> PAGE_SHIFT, data, kvm_read_guest_virt_helper()
4193 gpa_t gpa = vcpu->arch.walk_mmu->gva_to_gpa(vcpu, addr, access|PFERR_FETCH_MASK, kvm_fetch_guest_virt() local
4195 if (unlikely(gpa == UNMAPPED_GVA)) kvm_fetch_guest_virt()
4201 ret = kvm_vcpu_read_guest_page(vcpu, gpa >> PAGE_SHIFT, val, kvm_fetch_guest_virt()
4248 gpa_t gpa = vcpu->arch.walk_mmu->gva_to_gpa(vcpu, addr, kvm_write_guest_virt_system() local
4255 if (gpa == UNMAPPED_GVA) kvm_write_guest_virt_system()
4257 ret = kvm_vcpu_write_guest(vcpu, gpa, data, towrite); kvm_write_guest_virt_system()
4273 gpa_t *gpa, struct x86_exception *exception, vcpu_mmio_gva_to_gpa()
4282 *gpa = vcpu->arch.mmio_gfn << PAGE_SHIFT | vcpu_mmio_gva_to_gpa()
4284 trace_vcpu_match_mmio(gva, *gpa, write, false); vcpu_mmio_gva_to_gpa()
4288 *gpa = vcpu->arch.walk_mmu->gva_to_gpa(vcpu, gva, access, exception); vcpu_mmio_gva_to_gpa()
4290 if (*gpa == UNMAPPED_GVA) vcpu_mmio_gva_to_gpa()
4294 if ((*gpa & PAGE_MASK) == APIC_DEFAULT_PHYS_BASE) vcpu_mmio_gva_to_gpa()
4297 if (vcpu_match_mmio_gpa(vcpu, *gpa)) { vcpu_mmio_gva_to_gpa()
4298 trace_vcpu_match_mmio(gva, *gpa, write, true); vcpu_mmio_gva_to_gpa()
4305 int emulator_write_phys(struct kvm_vcpu *vcpu, gpa_t gpa, emulator_write_phys() argument
4310 ret = kvm_vcpu_write_guest(vcpu, gpa, val, bytes); emulator_write_phys()
4313 kvm_mmu_pte_write(vcpu, gpa, val, bytes); emulator_write_phys()
4320 int (*read_write_emulate)(struct kvm_vcpu *vcpu, gpa_t gpa,
4322 int (*read_write_mmio)(struct kvm_vcpu *vcpu, gpa_t gpa,
4324 int (*read_write_exit_mmio)(struct kvm_vcpu *vcpu, gpa_t gpa,
4333 vcpu->mmio_fragments[0].gpa, *(u64 *)val); read_prepare()
4341 static int read_emulate(struct kvm_vcpu *vcpu, gpa_t gpa, read_emulate() argument
4344 return !kvm_vcpu_read_guest(vcpu, gpa, val, bytes); read_emulate()
4347 static int write_emulate(struct kvm_vcpu *vcpu, gpa_t gpa, write_emulate() argument
4350 return emulator_write_phys(vcpu, gpa, val, bytes); write_emulate()
4353 static int write_mmio(struct kvm_vcpu *vcpu, gpa_t gpa, int bytes, void *val) write_mmio() argument
4355 trace_kvm_mmio(KVM_TRACE_MMIO_WRITE, bytes, gpa, *(u64 *)val); write_mmio()
4356 return vcpu_mmio_write(vcpu, gpa, bytes, val); write_mmio()
4359 static int read_exit_mmio(struct kvm_vcpu *vcpu, gpa_t gpa, read_exit_mmio() argument
4362 trace_kvm_mmio(KVM_TRACE_MMIO_READ_UNSATISFIED, bytes, gpa, 0); read_exit_mmio()
4366 static int write_exit_mmio(struct kvm_vcpu *vcpu, gpa_t gpa, write_exit_mmio() argument
4395 gpa_t gpa; emulator_read_write_onepage() local
4400 ret = vcpu_mmio_gva_to_gpa(vcpu, addr, &gpa, exception, write); emulator_read_write_onepage()
4409 if (ops->read_write_emulate(vcpu, gpa, val, bytes)) emulator_read_write_onepage()
4416 handled = ops->read_write_mmio(vcpu, gpa, bytes, val); emulator_read_write_onepage()
4420 gpa += handled; emulator_read_write_onepage()
4426 frag->gpa = gpa; emulator_read_write_onepage()
4439 gpa_t gpa; emulator_read_write() local
4473 gpa = vcpu->mmio_fragments[0].gpa; emulator_read_write()
4481 vcpu->run->mmio.phys_addr = gpa; emulator_read_write()
4483 return ops->read_write_exit_mmio(vcpu, gpa, val, bytes); emulator_read_write()
4524 gpa_t gpa; emulator_cmpxchg_emulated() local
4533 gpa = kvm_mmu_gva_to_gpa_write(vcpu, addr, NULL); emulator_cmpxchg_emulated()
4535 if (gpa == UNMAPPED_GVA || emulator_cmpxchg_emulated()
4536 (gpa & PAGE_MASK) == APIC_DEFAULT_PHYS_BASE) emulator_cmpxchg_emulated()
4539 if (((gpa + bytes - 1) & PAGE_MASK) != (gpa & PAGE_MASK)) emulator_cmpxchg_emulated()
4542 page = kvm_vcpu_gfn_to_page(vcpu, gpa >> PAGE_SHIFT); emulator_cmpxchg_emulated()
4547 kaddr += offset_in_page(gpa); emulator_cmpxchg_emulated()
4570 kvm_vcpu_mark_page_dirty(vcpu, gpa >> PAGE_SHIFT); emulator_cmpxchg_emulated()
4571 kvm_mmu_pte_write(vcpu, gpa, new, bytes); emulator_cmpxchg_emulated()
5114 gpa_t gpa = cr2; reexecute_instruction() local
5125 gpa = kvm_mmu_gva_to_gpa_write(vcpu, cr2, NULL); reexecute_instruction()
5131 if (gpa == UNMAPPED_GVA) reexecute_instruction()
5141 pfn = gfn_to_pfn(vcpu->kvm, gpa_to_gfn(gpa)); reexecute_instruction()
5161 kvm_mmu_unprotect_page(vcpu->kvm, gpa_to_gfn(gpa)); reexecute_instruction()
5171 kvm_mmu_unprotect_page(vcpu->kvm, gpa_to_gfn(gpa)); reexecute_instruction()
5185 unsigned long last_retry_eip, last_retry_addr, gpa = cr2; retry_instruction() local
5218 gpa = kvm_mmu_gva_to_gpa_write(vcpu, cr2, NULL); retry_instruction()
5220 kvm_mmu_unprotect_page(vcpu->kvm, gpa_to_gfn(gpa)); retry_instruction()
6733 * write gpa, len
6741 * write gpa, len
6766 frag->gpa += len; complete_emulated_mmio()
6781 run->mmio.phys_addr = frag->gpa; complete_emulated_mmio()
7151 gpa_t gpa; kvm_arch_vcpu_ioctl_translate() local
7155 gpa = kvm_mmu_gva_to_gpa_system(vcpu, vaddr, NULL); kvm_arch_vcpu_ioctl_translate()
7157 tr->physical_address = gpa; kvm_arch_vcpu_ioctl_translate()
7158 tr->valid = gpa != UNMAPPED_GVA; kvm_arch_vcpu_ioctl_translate()
7696 int __x86_set_memory_region(struct kvm *kvm, int id, gpa_t gpa, u32 size) __x86_set_memory_region() argument
7733 m.guest_phys_addr = gpa; __x86_set_memory_region()
7750 int x86_set_memory_region(struct kvm *kvm, int id, gpa_t gpa, u32 size) x86_set_memory_region() argument
7755 r = __x86_set_memory_region(kvm, id, gpa, size); x86_set_memory_region()
4272 vcpu_mmio_gva_to_gpa(struct kvm_vcpu *vcpu, unsigned long gva, gpa_t *gpa, struct x86_exception *exception, bool write) vcpu_mmio_gva_to_gpa() argument
H A Dpaging_tmpl.h878 gpa_t gpa = UNMAPPED_GVA; gva_to_gpa() local
884 gpa = gfn_to_gpa(walker.gfn); gva_to_gpa()
885 gpa |= vaddr & ~PAGE_MASK; gva_to_gpa()
889 return gpa; gva_to_gpa()
898 gpa_t gpa = UNMAPPED_GVA; gva_to_gpa_nested() local
904 gpa = gfn_to_gpa(walker.gfn); gva_to_gpa_nested()
905 gpa |= vaddr & ~PAGE_MASK; gva_to_gpa_nested()
909 return gpa; gva_to_gpa_nested()
H A Dmmu.c3481 static int tdp_page_fault(struct kvm_vcpu *vcpu, gva_t gpa, u32 error_code, tdp_page_fault() argument
3488 gfn_t gfn = gpa >> PAGE_SHIFT; tdp_page_fault()
3496 r = handle_mmio_page_fault(vcpu, gpa, true); tdp_page_fault()
3516 if (fast_page_fault(vcpu, gpa, level, error_code)) tdp_page_fault()
3522 if (try_async_pf(vcpu, prefault, gfn, gpa, &pfn, write, &map_writable)) tdp_page_fault()
3534 r = __direct_map(vcpu, gpa, write, map_writable, tdp_page_fault()
4173 static u64 mmu_pte_write_fetch_gpte(struct kvm_vcpu *vcpu, gpa_t *gpa, mmu_pte_write_fetch_gpte() argument
4186 *gpa &= ~(gpa_t)7; mmu_pte_write_fetch_gpte()
4188 r = kvm_vcpu_read_guest(vcpu, *gpa, &gentry, 8); mmu_pte_write_fetch_gpte()
4229 static bool detect_write_misaligned(struct kvm_mmu_page *sp, gpa_t gpa, detect_write_misaligned() argument
4234 pgprintk("misaligned: gpa %llx bytes %d role %x\n", detect_write_misaligned()
4235 gpa, bytes, sp->role.word); detect_write_misaligned()
4237 offset = offset_in_page(gpa); detect_write_misaligned()
4253 static u64 *get_written_sptes(struct kvm_mmu_page *sp, gpa_t gpa, int *nspte) get_written_sptes() argument
4259 page_offset = offset_in_page(gpa); get_written_sptes()
4284 void kvm_mmu_pte_write(struct kvm_vcpu *vcpu, gpa_t gpa, kvm_mmu_pte_write() argument
4287 gfn_t gfn = gpa >> PAGE_SHIFT; kvm_mmu_pte_write()
4311 pgprintk("%s: gpa %llx bytes %d\n", __func__, gpa, bytes); kvm_mmu_pte_write()
4313 gentry = mmu_pte_write_fetch_gpte(vcpu, &gpa, new, &bytes); kvm_mmu_pte_write()
4327 if (detect_write_misaligned(sp, gpa, bytes) || kvm_mmu_pte_write()
4335 spte = get_written_sptes(sp, gpa, &npte); kvm_mmu_pte_write()
4360 gpa_t gpa; kvm_mmu_unprotect_page_virt() local
4366 gpa = kvm_mmu_gva_to_gpa_read(vcpu, gva, NULL); kvm_mmu_unprotect_page_virt()
4368 r = kvm_mmu_unprotect_page(vcpu->kvm, gpa >> PAGE_SHIFT); kvm_mmu_unprotect_page_virt()
H A Dvmx.c1286 static inline void __invept(int ext, u64 eptp, gpa_t gpa) __invept() argument
1289 u64 eptp, gpa; __invept() member in struct:__anon3202
1290 } operand = {eptp, gpa}; __invept()
5880 gpa_t gpa; handle_ept_violation() local
5910 gpa = vmcs_read64(GUEST_PHYSICAL_ADDRESS); handle_ept_violation()
5911 trace_kvm_page_fault(gpa, exit_qualification); handle_ept_violation()
5922 return kvm_mmu_page_fault(vcpu, gpa, error_code, NULL, 0); handle_ept_violation()
5928 gpa_t gpa; handle_ept_misconfig() local
5930 gpa = vmcs_read64(GUEST_PHYSICAL_ADDRESS); handle_ept_misconfig()
5931 if (!kvm_io_bus_write(vcpu, KVM_FAST_MMIO_BUS, gpa, 0, NULL)) { handle_ept_misconfig()
5933 trace_kvm_fast_mmio(gpa); handle_ept_misconfig()
5937 ret = handle_mmio_page_fault(vcpu, gpa, true); handle_ept_misconfig()
5939 return x86_emulate_instruction(vcpu, gpa, 0, NULL, 0) == handle_ept_misconfig()
5943 return kvm_mmu_page_fault(vcpu, gpa, 0, NULL, 0); handle_ept_misconfig()
7324 u64 eptp, gpa; handle_invept() member in struct:__anon3203
7886 u64 gpa; vmx_flush_pml_buffer() local
7888 gpa = pml_buf[pml_idx]; vmx_flush_pml_buffer()
7889 WARN_ON(gpa & (PAGE_SIZE - 1)); vmx_flush_pml_buffer()
7890 kvm_vcpu_mark_page_dirty(vcpu, gpa >> PAGE_SHIFT); vmx_flush_pml_buffer()
9382 static u32 nested_vmx_load_msr(struct kvm_vcpu *vcpu, u64 gpa, u32 count) nested_vmx_load_msr() argument
9390 if (kvm_vcpu_read_guest(vcpu, gpa + i * sizeof(e), nested_vmx_load_msr()
9394 __func__, i, gpa + i * sizeof(e)); nested_vmx_load_msr()
9417 static int nested_vmx_store_msr(struct kvm_vcpu *vcpu, u64 gpa, u32 count) nested_vmx_store_msr() argument
9425 gpa + i * sizeof(e), nested_vmx_store_msr()
9429 __func__, i, gpa + i * sizeof(e)); nested_vmx_store_msr()
9447 gpa + i * sizeof(e) + nested_vmx_store_msr()
H A Dsvm.c104 /* gpa pointers to the real vectors */
2028 static void *nested_svm_map(struct vcpu_svm *svm, u64 gpa, struct page **_page) nested_svm_map() argument
2034 page = kvm_vcpu_gfn_to_page(&svm->vcpu, gpa >> PAGE_SHIFT); nested_svm_map()
2059 u64 gpa; nested_svm_intercept_ioio() local
2067 gpa = svm->nested.vmcb_iopm + (port / 8); nested_svm_intercept_ioio()
2073 if (kvm_vcpu_read_guest(&svm->vcpu, gpa, &val, iopm_len)) nested_svm_intercept_ioio()
/linux-4.4.14/drivers/misc/sgi-xp/
H A Dxp_uv.c39 xp_socket_pa_uv(unsigned long gpa) xp_socket_pa_uv() argument
41 return uv_gpa_to_soc_phys_ram(gpa); xp_socket_pa_uv()
H A Dxp_sn2.c90 xp_socket_pa_sn2(unsigned long gpa) xp_socket_pa_sn2() argument
92 return gpa; xp_socket_pa_sn2()
H A Dxp_main.c47 unsigned long (*xp_socket_pa) (unsigned long gpa);
/linux-4.4.14/virt/kvm/
H A Dkvm_main.c1656 int kvm_read_guest(struct kvm *kvm, gpa_t gpa, void *data, unsigned long len) kvm_read_guest() argument
1658 gfn_t gfn = gpa >> PAGE_SHIFT; kvm_read_guest()
1660 int offset = offset_in_page(gpa); kvm_read_guest()
1676 int kvm_vcpu_read_guest(struct kvm_vcpu *vcpu, gpa_t gpa, void *data, unsigned long len) kvm_vcpu_read_guest() argument
1678 gfn_t gfn = gpa >> PAGE_SHIFT; kvm_vcpu_read_guest()
1680 int offset = offset_in_page(gpa); kvm_vcpu_read_guest()
1713 int kvm_read_guest_atomic(struct kvm *kvm, gpa_t gpa, void *data, kvm_read_guest_atomic() argument
1716 gfn_t gfn = gpa >> PAGE_SHIFT; kvm_read_guest_atomic()
1718 int offset = offset_in_page(gpa); kvm_read_guest_atomic()
1724 int kvm_vcpu_read_guest_atomic(struct kvm_vcpu *vcpu, gpa_t gpa, kvm_vcpu_read_guest_atomic() argument
1727 gfn_t gfn = gpa >> PAGE_SHIFT; kvm_vcpu_read_guest_atomic()
1729 int offset = offset_in_page(gpa); kvm_vcpu_read_guest_atomic()
1769 int kvm_write_guest(struct kvm *kvm, gpa_t gpa, const void *data, kvm_write_guest() argument
1772 gfn_t gfn = gpa >> PAGE_SHIFT; kvm_write_guest()
1774 int offset = offset_in_page(gpa); kvm_write_guest()
1790 int kvm_vcpu_write_guest(struct kvm_vcpu *vcpu, gpa_t gpa, const void *data, kvm_vcpu_write_guest() argument
1793 gfn_t gfn = gpa >> PAGE_SHIFT; kvm_vcpu_write_guest()
1795 int offset = offset_in_page(gpa); kvm_vcpu_write_guest()
1812 gpa_t gpa, unsigned long len) kvm_gfn_to_hva_cache_init()
1815 int offset = offset_in_page(gpa); kvm_gfn_to_hva_cache_init()
1816 gfn_t start_gfn = gpa >> PAGE_SHIFT; kvm_gfn_to_hva_cache_init()
1817 gfn_t end_gfn = (gpa + len - 1) >> PAGE_SHIFT; kvm_gfn_to_hva_cache_init()
1821 ghc->gpa = gpa; kvm_gfn_to_hva_cache_init()
1857 kvm_gfn_to_hva_cache_init(kvm, ghc, ghc->gpa, ghc->len); kvm_write_guest_cached()
1860 return kvm_write_guest(kvm, ghc->gpa, data, len); kvm_write_guest_cached()
1868 mark_page_dirty_in_slot(ghc->memslot, ghc->gpa >> PAGE_SHIFT); kvm_write_guest_cached()
1883 kvm_gfn_to_hva_cache_init(kvm, ghc, ghc->gpa, ghc->len); kvm_read_guest_cached()
1886 return kvm_read_guest(kvm, ghc->gpa, data, len); kvm_read_guest_cached()
1907 int kvm_clear_guest(struct kvm *kvm, gpa_t gpa, unsigned long len) kvm_clear_guest() argument
1909 gfn_t gfn = gpa >> PAGE_SHIFT; kvm_clear_guest()
1911 int offset = offset_in_page(gpa); kvm_clear_guest()
1811 kvm_gfn_to_hva_cache_init(struct kvm *kvm, struct gfn_to_hva_cache *ghc, gpa_t gpa, unsigned long len) kvm_gfn_to_hva_cache_init() argument
/linux-4.4.14/include/xen/interface/hvm/
H A Dhvm_op.h41 aligned_u64 gpa; member in struct:xen_hvm_pagetable_dying
/linux-4.4.14/drivers/misc/sgi-gru/
H A Dgrufault.c262 int write, int atomic, unsigned long *gpa, int *pageshift) gru_vtop()
288 *gpa = uv_soc_phys_ram_to_gpa(paddr); gru_vtop()
325 unsigned long vaddr = 0, gpa; gru_preload_tlb() local
341 ret = gru_vtop(gts, vaddr, write, atomic, &gpa, &pageshift); gru_preload_tlb()
342 if (ret || tfh_write_only(tfh, gpa, GAA_RAM, vaddr, asid, write, gru_preload_tlb()
346 "%s: gid %d, gts 0x%p, tfh 0x%p, vaddr 0x%lx, asid 0x%x, rw %d, ps %d, gpa 0x%lx\n", gru_preload_tlb()
348 vaddr, asid, write, pageshift, gpa); gru_preload_tlb()
372 unsigned long gpa = 0, vaddr = 0; gru_try_dropin() local
422 ret = gru_vtop(gts, vaddr, write, atomic, &gpa, &pageshift); gru_try_dropin()
443 tfh_write_restart(tfh, gpa, GAA_RAM, vaddr, asid, write, gru_try_dropin()
447 " rw %d, ps %d, gpa 0x%lx\n", gru_try_dropin()
449 indexway, write, pageshift, gpa); gru_try_dropin()
261 gru_vtop(struct gru_thread_state *gts, unsigned long vaddr, int write, int atomic, unsigned long *gpa, int *pageshift) gru_vtop() argument
H A Dgrukservices.h138 * gpa source numalink physical address of GPA
144 int gru_read_gpa(unsigned long *value, unsigned long gpa);
H A Dgru_instructions.h359 static inline void gru_vload_phys(void *cb, unsigned long gpa, gru_vload_phys() argument
364 ins->baddr0 = (long)gpa | ((unsigned long)iaa << 62); gru_vload_phys()
371 static inline void gru_vstore_phys(void *cb, unsigned long gpa, gru_vstore_phys() argument
376 ins->baddr0 = (long)gpa | ((unsigned long)iaa << 62); gru_vstore_phys()
H A Dgrukservices.c721 unsigned long m, *val = mesg, gpa, save; send_message_put_nacked() local
744 gpa = uv_global_gru_mmr_address(mqd->interrupt_pnode, UVH_IPI_INT); send_message_put_nacked()
748 gru_vstore_phys(cb, gpa, gru_get_tri(mesg), IAA_REGISTER, IMA); send_message_put_nacked()
911 int gru_read_gpa(unsigned long *value, unsigned long gpa) gru_read_gpa() argument
920 iaa = gpa >> 62; gru_read_gpa()
921 gru_vload_phys(cb, gpa, gru_get_tri(dsr), iaa, IMA); gru_read_gpa()
/linux-4.4.14/arch/powerpc/kvm/
H A De500.h258 gpa_t gpa; tlbe_is_host_safe() local
270 gpa = get_tlb_raddr(tlbe); tlbe_is_host_safe()
271 if (!gfn_to_memslot(vcpu->kvm, gpa >> PAGE_SHIFT)) tlbe_is_host_safe()
H A Dbook3s_64_mmu_hv.c392 unsigned long gpa, gva_t ea, int is_store) kvmppc_hv_emulate_mmio()
431 vcpu->arch.paddr_accessed = gpa; kvmppc_hv_emulate_mmio()
444 unsigned long gpa, gfn, hva, pfn; kvmppc_book3s_hv_page_fault() local
483 gpa = gpa_base | (ea & (psize - 1)); kvmppc_book3s_hv_page_fault()
484 gfn = gpa >> PAGE_SHIFT; kvmppc_book3s_hv_page_fault()
491 return kvmppc_hv_emulate_mmio(run, vcpu, gpa, ea, kvmppc_book3s_hv_page_fault()
1018 gfn = vpa->gpa >> PAGE_SHIFT; harvest_vpa_dirty()
1062 void *kvmppc_pin_guest_page(struct kvm *kvm, unsigned long gpa, kvmppc_pin_guest_page() argument
1066 unsigned long gfn = gpa >> PAGE_SHIFT; kvmppc_pin_guest_page()
1083 offset = gpa & (PAGE_SIZE - 1); kvmppc_pin_guest_page()
1093 void kvmppc_unpin_guest_page(struct kvm *kvm, void *va, unsigned long gpa, kvmppc_unpin_guest_page() argument
1108 gfn = gpa >> PAGE_SHIFT; kvmppc_unpin_guest_page()
391 kvmppc_hv_emulate_mmio(struct kvm_run *run, struct kvm_vcpu *vcpu, unsigned long gpa, gva_t ea, int is_store) kvmppc_hv_emulate_mmio() argument
H A Dbook3s.c369 pfn_t kvmppc_gpa_to_pfn(struct kvm_vcpu *vcpu, gpa_t gpa, bool writing, kvmppc_gpa_to_pfn() argument
373 gfn_t gfn = gpa >> PAGE_SHIFT; kvmppc_gpa_to_pfn()
379 gpa &= ~0xFFFULL; kvmppc_gpa_to_pfn()
380 if (unlikely(mp_pa) && unlikely((gpa & KVM_PAM) == mp_pa)) { kvmppc_gpa_to_pfn()
H A Dbook3s_32_mmu_host.c159 /* Get host physical address for gpa */ kvmppc_mmu_map_page()
162 printk(KERN_INFO "Couldn't get guest page for gpa %lx!\n", kvmppc_mmu_map_page()
H A Dbook3s_64_mmu_host.c107 /* Get host physical address for gpa */ kvmppc_mmu_map_page()
110 printk(KERN_INFO "Couldn't get guest page for gpa %lx!\n", kvmppc_mmu_map_page()
H A Dbook3s_hv.c484 unsigned long gpa; kvmppc_update_vpa() local
495 gpa = vpap->next_gpa; kvmppc_update_vpa()
499 if (gpa) kvmppc_update_vpa()
500 va = kvmppc_pin_guest_page(kvm, gpa, &nb); kvmppc_update_vpa()
502 if (gpa == vpap->next_gpa) kvmppc_update_vpa()
506 kvmppc_unpin_guest_page(kvm, va, gpa, false); kvmppc_update_vpa()
516 kvmppc_unpin_guest_page(kvm, va, gpa, false); kvmppc_update_vpa()
520 kvmppc_unpin_guest_page(kvm, vpap->pinned_addr, vpap->gpa, kvmppc_update_vpa()
522 vpap->gpa = gpa; kvmppc_update_vpa()
1691 kvmppc_unpin_guest_page(kvm, vpa->pinned_addr, vpa->gpa, unpin_vpa()
H A Dbook3s_hv_rm_mmu.c171 unsigned long i, pa, gpa, gfn, psize; kvmppc_do_h_enter() local
198 gpa = (ptel & HPTE_R_RPN) & ~(psize - 1); kvmppc_do_h_enter()
199 gfn = gpa >> PAGE_SHIFT; kvmppc_do_h_enter()
256 pa |= gpa & ~PAGE_MASK; kvmppc_do_h_enter()
H A Dbook3s_pr.c515 static int kvmppc_visible_gpa(struct kvm_vcpu *vcpu, gpa_t gpa) kvmppc_visible_gpa() argument
522 gpa &= ~0xFFFULL; kvmppc_visible_gpa()
523 if (unlikely(mp_pa) && unlikely((mp_pa & KVM_PAM) == (gpa & KVM_PAM))) { kvmppc_visible_gpa()
527 return kvm_is_visible_gfn(vcpu->kvm, gpa >> PAGE_SHIFT); kvmppc_visible_gpa()
/linux-4.4.14/arch/arm/kvm/
H A Dmmu.c778 gpa_t gpa = addr + (vm_start - memslot->userspace_addr); stage2_unmap_memslot() local
779 unmap_stage2_range(kvm, gpa, vm_end - vm_start); stage2_unmap_memslot()
1470 gpa_t gpa, void *data), handle_hva_to_gpa()
1498 gpa_t gpa = gfn << PAGE_SHIFT; kvm_for_each_memslot() local
1499 ret |= handler(kvm, gpa, data); kvm_for_each_memslot()
1506 static int kvm_unmap_hva_handler(struct kvm *kvm, gpa_t gpa, void *data) kvm_unmap_hva_handler() argument
1508 unmap_stage2_range(kvm, gpa, PAGE_SIZE); kvm_unmap_hva_handler()
1535 static int kvm_set_spte_handler(struct kvm *kvm, gpa_t gpa, void *data) kvm_set_spte_handler() argument
1546 stage2_set_pte(kvm, NULL, gpa, pte, 0); kvm_set_spte_handler()
1564 static int kvm_age_hva_handler(struct kvm *kvm, gpa_t gpa, void *data) kvm_age_hva_handler() argument
1569 pmd = stage2_get_pmd(kvm, NULL, gpa); kvm_age_hva_handler()
1582 pte = pte_offset_kernel(pmd, gpa); kvm_age_hva_handler()
1594 static int kvm_test_age_hva_handler(struct kvm *kvm, gpa_t gpa, void *data) kvm_test_age_hva_handler() argument
1599 pmd = stage2_get_pmd(kvm, NULL, gpa); kvm_test_age_hva_handler()
1606 pte = pte_offset_kernel(pmd, gpa); kvm_test_age_hva_handler()
1797 gpa_t gpa = mem->guest_phys_addr + kvm_arch_prepare_memory_region() local
1808 ret = kvm_phys_addr_ioremap(kvm, gpa, pa, kvm_arch_prepare_memory_region()
1860 gpa_t gpa = slot->base_gfn << PAGE_SHIFT; kvm_arch_flush_shadow_memslot() local
1864 unmap_stage2_range(kvm, gpa, size); kvm_arch_flush_shadow_memslot()
1466 handle_hva_to_gpa(struct kvm *kvm, unsigned long start, unsigned long end, int (*handler)(struct kvm *kvm, gpa_t gpa, void *data), void *data) handle_hva_to_gpa() argument
/linux-4.4.14/arch/mips/kvm/
H A Dtrap_emul.c24 gpa_t gpa; kvm_trap_emul_gva_to_gpa_cb() local
28 gpa = CPHYSADDR(gva); kvm_trap_emul_gva_to_gpa_cb()
32 gpa = KVM_INVALID_ADDR; kvm_trap_emul_gva_to_gpa_cb()
35 kvm_debug("%s: gva %#lx, gpa: %#llx\n", __func__, gva, gpa); kvm_trap_emul_gva_to_gpa_cb()
37 return gpa; kvm_trap_emul_gva_to_gpa_cb()
/linux-4.4.14/drivers/pinctrl/samsung/
H A Dpinctrl-s3c24xx.c562 PIN_BANK_A(23, 0x000, "gpa"),
582 PIN_BANK_A(27, 0x000, "gpa"),
604 PIN_BANK_A(25, 0x000, "gpa"),
624 PIN_BANK_A(28, 0x000, "gpa"),
H A Dpinctrl-s3c64xx.c780 PIN_BANK_4BIT_EINTG(8, 0x000, "gpa", 0),
/linux-4.4.14/tools/perf/arch/x86/util/
H A Dkvm-stat.c22 key->key = perf_evsel__intval(evsel, sample, "gpa"); mmio_event_get_key()
/linux-4.4.14/arch/x86/include/asm/
H A Dkvm_host.h279 gpa_t (*translate_gpa)(struct kvm_vcpu *vcpu, gpa_t gpa, u32 access,
955 int emulator_write_phys(struct kvm_vcpu *vcpu, gpa_t gpa,
1070 void kvm_mmu_pte_write(struct kvm_vcpu *vcpu, gpa_t gpa,
1078 gpa_t translate_nested_gpa(struct kvm_vcpu *vcpu, gpa_t gpa, u32 access,
1099 static inline gpa_t translate_gpa(struct kvm_vcpu *vcpu, gpa_t gpa, u32 access, translate_gpa() argument
1102 return gpa; translate_gpa()
1247 int __x86_set_memory_region(struct kvm *kvm, int id, gpa_t gpa, u32 size);
1248 int x86_set_memory_region(struct kvm *kvm, int id, gpa_t gpa, u32 size);
H A Dkvm_emulate.h25 u64 address; /* cr2 or nested page fault gpa */
/linux-4.4.14/arch/ia64/include/asm/uv/
H A Duv_hub.h166 static inline void *uv_va(unsigned long gpa) uv_va() argument
168 return __va(gpa & uv_hub_info->gpa_mask); uv_va()
/linux-4.4.14/arch/powerpc/include/asm/
H A Dkvm_book3s.h157 extern pfn_t kvmppc_gpa_to_pfn(struct kvm_vcpu *vcpu, gpa_t gpa, bool writing,
169 unsigned long gpa, bool dirty);
H A Dkvm_host.h329 unsigned long gpa; /* Current guest phys addr */ member in struct:kvmppc_vpa
/linux-4.4.14/arch/x86/platform/uv/
H A Dtlb_uv.c1693 unsigned long gpa; activation_descriptor_init() local
1711 gpa = uv_gpa(bau_desc); activation_descriptor_init()
1712 n = uv_gpa_to_gnode(gpa); activation_descriptor_init()
1713 m = uv_gpa_to_offset(gpa); activation_descriptor_init()
/linux-4.4.14/kernel/rcu/
H A Dtree.c1194 unsigned long gpa; rcu_check_gp_kthread_starvation() local
1198 gpa = READ_ONCE(rsp->gp_activity); rcu_check_gp_kthread_starvation()
1199 if (j - gpa > 2 * HZ) rcu_check_gp_kthread_starvation()
1201 rsp->name, j - gpa, rcu_check_gp_kthread_starvation()
1232 unsigned long gpa; print_other_cpu_stall() local
1286 gpa = READ_ONCE(rsp->gp_activity);
1288 rsp->name, j - gpa, j, gpa,
/linux-4.4.14/arch/x86/xen/
H A Dmmu.c2751 a.gpa = __pa(mm->pgd); xen_hvm_exit_mmap()
2762 a.gpa = 0x00; is_pagetable_dying_supported()

Completed in 1637 milliseconds