/linux-4.4.14/include/linux/ |
H A D | kvm_types.h | 41 * gpa - guest physical address 60 gpa_t gpa; member in struct:gfn_to_hva_cache
|
H A D | kvm_host.h | 220 gpa_t gpa; member in struct:kvm_mmio_fragment 621 int kvm_read_guest_atomic(struct kvm *kvm, gpa_t gpa, void *data, 623 int kvm_read_guest(struct kvm *kvm, gpa_t gpa, void *data, unsigned long len); 628 int kvm_write_guest(struct kvm *kvm, gpa_t gpa, const void *data, 633 gpa_t gpa, unsigned long len); 635 int kvm_clear_guest(struct kvm *kvm, gpa_t gpa, unsigned long len); 650 int kvm_vcpu_read_guest_atomic(struct kvm_vcpu *vcpu, gpa_t gpa, void *data, 652 int kvm_vcpu_read_guest(struct kvm_vcpu *vcpu, gpa_t gpa, void *data, 656 int kvm_vcpu_write_guest(struct kvm_vcpu *vcpu, gpa_t gpa, const void *data, 976 static inline gfn_t gpa_to_gfn(gpa_t gpa) gpa_to_gfn() argument 978 return (gfn_t)(gpa >> PAGE_SHIFT); gpa_to_gfn() 986 static inline bool kvm_is_error_gpa(struct kvm *kvm, gpa_t gpa) kvm_is_error_gpa() argument 988 unsigned long hva = gfn_to_hva(kvm, gpa_to_gfn(gpa)); kvm_is_error_gpa()
|
/linux-4.4.14/arch/s390/kvm/ |
H A D | gaccess.h | 127 unsigned long gpa = gra + kvm_s390_get_prefix(vcpu); write_guest_lc() local 129 return kvm_write_guest(vcpu->kvm, gpa, data, len); write_guest_lc() 153 unsigned long gpa = gra + kvm_s390_get_prefix(vcpu); read_guest_lc() local 155 return kvm_read_guest(vcpu->kvm, gpa, data, len); read_guest_lc() 159 ar_t ar, unsigned long *gpa, int write); 244 * @gpa: guest physical (absolute) address 248 * Copy @len bytes from @data (kernel space) to @gpa (guest absolute address). 258 int write_guest_abs(struct kvm_vcpu *vcpu, unsigned long gpa, void *data, write_guest_abs() argument 261 return kvm_write_guest(vcpu->kvm, gpa, data, len); write_guest_abs() 267 * @gpa: guest physical (absolute) address 271 * Copy @len bytes from @gpa (guest absolute address) to @data (kernel space). 281 int read_guest_abs(struct kvm_vcpu *vcpu, unsigned long gpa, void *data, read_guest_abs() argument 284 return kvm_read_guest(vcpu->kvm, gpa, data, len); read_guest_abs()
|
H A D | gaccess.c | 514 static int deref_table(struct kvm *kvm, unsigned long gpa, unsigned long *val) deref_table() argument 516 return kvm_read_guest(kvm, gpa, val, sizeof(*val)); deref_table() 523 * @gpa: points to where guest physical (absolute) address should be stored 530 * an addressing exception is indicated and @gpa will not be changed. 532 * Returns: - zero on success; @gpa contains the resulting absolute address 540 unsigned long *gpa, const union asce asce, guest_translate() 690 *gpa = raddr.addr; guest_translate() 758 unsigned long _len, nr_pages, gpa, idx; access_guest() local 781 gpa = *(pages + idx) + (ga & ~PAGE_MASK); access_guest() 782 _len = min(PAGE_SIZE - (gpa & ~PAGE_MASK), len); access_guest() 784 rc = kvm_write_guest(vcpu->kvm, gpa, data, _len); access_guest() 786 rc = kvm_read_guest(vcpu->kvm, gpa, data, _len); access_guest() 801 unsigned long _len, gpa; access_guest_real() local 805 gpa = kvm_s390_real_to_abs(vcpu, gra); access_guest_real() 806 _len = min(PAGE_SIZE - (gpa & ~PAGE_MASK), len); access_guest_real() 808 rc = write_guest_abs(vcpu, gpa, data, _len); access_guest_real() 810 rc = read_guest_abs(vcpu, gpa, data, _len); access_guest_real() 828 unsigned long *gpa, int write) guest_translate_address() 850 rc = guest_translate(vcpu, gva, gpa, asce, write); guest_translate_address() 858 *gpa = kvm_s390_real_to_abs(vcpu, gva); guest_translate_address() 859 if (kvm_is_error_gpa(vcpu->kvm, *gpa)) guest_translate_address() 872 unsigned long gpa; check_gva_range() local 879 rc = guest_translate_address(vcpu, gva, ar, &gpa, is_write); check_gva_range() 539 guest_translate(struct kvm_vcpu *vcpu, unsigned long gva, unsigned long *gpa, const union asce asce, int write) guest_translate() argument 827 guest_translate_address(struct kvm_vcpu *vcpu, unsigned long gva, ar_t ar, unsigned long *gpa, int write) guest_translate_address() argument
|
H A D | kvm-s390.c | 1881 * @gpa: Guest physical address 1888 long kvm_arch_fault_in_page(struct kvm_vcpu *vcpu, gpa_t gpa, int writable) kvm_arch_fault_in_page() argument 1890 return gmap_fault(vcpu->arch.gmap, gpa, kvm_arch_fault_in_page() 2062 gpa_t gpa = current->thread.gmap_addr; vcpu_post_run() local 2063 rc = kvm_arch_fault_in_page(vcpu, gpa, 1); vcpu_post_run() 2230 int kvm_s390_store_status_unloaded(struct kvm_vcpu *vcpu, unsigned long gpa) kvm_s390_store_status_unloaded() argument 2239 if (gpa == KVM_S390_STORE_STATUS_NOADDR) { kvm_s390_store_status_unloaded() 2242 gpa = 0; kvm_s390_store_status_unloaded() 2243 } else if (gpa == KVM_S390_STORE_STATUS_PREFIXED) { kvm_s390_store_status_unloaded() 2246 gpa = px; kvm_s390_store_status_unloaded() 2248 gpa -= __LC_FPREGS_SAVE_AREA; kvm_s390_store_status_unloaded() 2253 rc = write_guest_abs(vcpu, gpa + __LC_FPREGS_SAVE_AREA, kvm_s390_store_status_unloaded() 2256 rc = write_guest_abs(vcpu, gpa + __LC_FPREGS_SAVE_AREA, kvm_s390_store_status_unloaded() 2259 rc |= write_guest_abs(vcpu, gpa + __LC_GPREGS_SAVE_AREA, kvm_s390_store_status_unloaded() 2261 rc |= write_guest_abs(vcpu, gpa + __LC_PSW_SAVE_AREA, kvm_s390_store_status_unloaded() 2263 rc |= write_guest_abs(vcpu, gpa + __LC_PREFIX_SAVE_AREA, kvm_s390_store_status_unloaded() 2265 rc |= write_guest_abs(vcpu, gpa + __LC_FP_CREG_SAVE_AREA, kvm_s390_store_status_unloaded() 2267 rc |= write_guest_abs(vcpu, gpa + __LC_TOD_PROGREG_SAVE_AREA, kvm_s390_store_status_unloaded() 2269 rc |= write_guest_abs(vcpu, gpa + __LC_CPU_TIMER_SAVE_AREA, kvm_s390_store_status_unloaded() 2272 rc |= write_guest_abs(vcpu, gpa + __LC_CLOCK_COMP_SAVE_AREA, kvm_s390_store_status_unloaded() 2274 rc |= write_guest_abs(vcpu, gpa + __LC_AREGS_SAVE_AREA, kvm_s390_store_status_unloaded() 2276 rc |= write_guest_abs(vcpu, gpa + __LC_CREGS_SAVE_AREA, kvm_s390_store_status_unloaded() 2299 unsigned long gpa) kvm_s390_store_adtl_status_unloaded() 2302 if (!(gpa & ~0x3ff)) kvm_s390_store_adtl_status_unloaded() 2305 return write_guest_abs(vcpu, gpa & ~0x3ff, kvm_s390_store_adtl_status_unloaded() 2298 kvm_s390_store_adtl_status_unloaded(struct kvm_vcpu *vcpu, unsigned long gpa) kvm_s390_store_adtl_status_unloaded() argument
|
H A D | priv.c | 965 unsigned long hva, gpa; handle_tprot() local 984 ret = guest_translate_address(vcpu, address1, ar, &gpa, 1); handle_tprot() 988 ret = guest_translate_address(vcpu, address1, ar, &gpa, 0); handle_tprot() 1001 hva = gfn_to_hva_prot(vcpu->kvm, gpa_to_gfn(gpa), &writable); handle_tprot()
|
H A D | kvm-s390.h | 235 long kvm_arch_fault_in_page(struct kvm_vcpu *vcpu, gpa_t gpa, int writable);
|
/linux-4.4.14/include/trace/events/ |
H A D | kvm.h | 207 TP_PROTO(int type, int len, u64 gpa, u64 val), 208 TP_ARGS(type, len, gpa, val), 213 __field( u64, gpa ) 220 __entry->gpa = gpa; 224 TP_printk("mmio %s len %u gpa 0x%llx val 0x%llx", 226 __entry->len, __entry->gpa, __entry->val)
|
/linux-4.4.14/arch/x86/include/asm/uv/ |
H A D | uv_hub.h | 330 uv_gpa_in_mmr_space(unsigned long gpa) uv_gpa_in_mmr_space() argument 332 return (gpa >> 62) == 0x3UL; uv_gpa_in_mmr_space() 336 static inline unsigned long uv_gpa_to_soc_phys_ram(unsigned long gpa) uv_gpa_to_soc_phys_ram() argument 342 gpa = ((gpa << uv_hub_info->m_shift) >> uv_hub_info->m_shift) | uv_gpa_to_soc_phys_ram() 343 ((gpa >> uv_hub_info->n_lshift) << uv_hub_info->m_val); uv_gpa_to_soc_phys_ram() 344 paddr = gpa & uv_hub_info->gpa_mask; uv_gpa_to_soc_phys_ram() 351 /* gpa -> pnode */ uv_gpa_to_gnode() 352 static inline unsigned long uv_gpa_to_gnode(unsigned long gpa) uv_gpa_to_gnode() argument 354 return gpa >> uv_hub_info->n_lshift; uv_gpa_to_gnode() 357 /* gpa -> pnode */ uv_gpa_to_pnode() 358 static inline int uv_gpa_to_pnode(unsigned long gpa) uv_gpa_to_pnode() argument 362 return uv_gpa_to_gnode(gpa) & n_mask; uv_gpa_to_pnode() 365 /* gpa -> node offset*/ uv_gpa_to_offset() 366 static inline unsigned long uv_gpa_to_offset(unsigned long gpa) uv_gpa_to_offset() argument 368 return (gpa << uv_hub_info->m_shift) >> uv_hub_info->m_shift; uv_gpa_to_offset()
|
/linux-4.4.14/arch/x86/kvm/ |
H A D | trace.h | 135 TP_PROTO(u64 gpa), 136 TP_ARGS(gpa), 139 __field(u64, gpa) 143 __entry->gpa = gpa; 146 TP_printk("fast mmio at gpa 0x%llx", __entry->gpa) 765 TP_PROTO(gva_t gva, gpa_t gpa, bool write, bool gpa_match), 766 TP_ARGS(gva, gpa, write, gpa_match), 770 __field(gpa_t, gpa) 777 __entry->gpa = gpa; 782 TP_printk("gva %#lx gpa %#llx %s %s", __entry->gva, __entry->gpa,
|
H A D | mmutrace.h | 103 __field(__u64, gpa) 107 __entry->gpa = ((u64)table_gfn << PAGE_SHIFT) 111 TP_printk("gpa %llx", __entry->gpa)
|
H A D | x86.h | 124 static inline bool vcpu_match_mmio_gpa(struct kvm_vcpu *vcpu, gpa_t gpa) vcpu_match_mmio_gpa() argument 127 vcpu->arch.mmio_gfn == gpa >> PAGE_SHIFT) vcpu_match_mmio_gpa()
|
H A D | x86.c | 1968 gpa_t gpa = data & ~0x3f; kvm_pv_enable_async_pf() local 1982 if (kvm_gfn_to_hva_cache_init(vcpu->kvm, &vcpu->arch.apf.data, gpa, kvm_pv_enable_async_pf() 4107 gpa_t translate_nested_gpa(struct kvm_vcpu *vcpu, gpa_t gpa, u32 access, translate_nested_gpa() argument 4116 t_gpa = vcpu->arch.mmu.gva_to_gpa(vcpu, gpa, access, exception); translate_nested_gpa() 4159 gpa_t gpa = vcpu->arch.walk_mmu->gva_to_gpa(vcpu, addr, access, kvm_read_guest_virt_helper() local 4165 if (gpa == UNMAPPED_GVA) kvm_read_guest_virt_helper() 4167 ret = kvm_vcpu_read_guest_page(vcpu, gpa >> PAGE_SHIFT, data, kvm_read_guest_virt_helper() 4193 gpa_t gpa = vcpu->arch.walk_mmu->gva_to_gpa(vcpu, addr, access|PFERR_FETCH_MASK, kvm_fetch_guest_virt() local 4195 if (unlikely(gpa == UNMAPPED_GVA)) kvm_fetch_guest_virt() 4201 ret = kvm_vcpu_read_guest_page(vcpu, gpa >> PAGE_SHIFT, val, kvm_fetch_guest_virt() 4248 gpa_t gpa = vcpu->arch.walk_mmu->gva_to_gpa(vcpu, addr, kvm_write_guest_virt_system() local 4255 if (gpa == UNMAPPED_GVA) kvm_write_guest_virt_system() 4257 ret = kvm_vcpu_write_guest(vcpu, gpa, data, towrite); kvm_write_guest_virt_system() 4273 gpa_t *gpa, struct x86_exception *exception, vcpu_mmio_gva_to_gpa() 4282 *gpa = vcpu->arch.mmio_gfn << PAGE_SHIFT | vcpu_mmio_gva_to_gpa() 4284 trace_vcpu_match_mmio(gva, *gpa, write, false); vcpu_mmio_gva_to_gpa() 4288 *gpa = vcpu->arch.walk_mmu->gva_to_gpa(vcpu, gva, access, exception); vcpu_mmio_gva_to_gpa() 4290 if (*gpa == UNMAPPED_GVA) vcpu_mmio_gva_to_gpa() 4294 if ((*gpa & PAGE_MASK) == APIC_DEFAULT_PHYS_BASE) vcpu_mmio_gva_to_gpa() 4297 if (vcpu_match_mmio_gpa(vcpu, *gpa)) { vcpu_mmio_gva_to_gpa() 4298 trace_vcpu_match_mmio(gva, *gpa, write, true); vcpu_mmio_gva_to_gpa() 4305 int emulator_write_phys(struct kvm_vcpu *vcpu, gpa_t gpa, emulator_write_phys() argument 4310 ret = kvm_vcpu_write_guest(vcpu, gpa, val, bytes); emulator_write_phys() 4313 kvm_mmu_pte_write(vcpu, gpa, val, bytes); emulator_write_phys() 4320 int (*read_write_emulate)(struct kvm_vcpu *vcpu, gpa_t gpa, 4322 int (*read_write_mmio)(struct kvm_vcpu *vcpu, gpa_t gpa, 4324 int (*read_write_exit_mmio)(struct kvm_vcpu *vcpu, gpa_t gpa, 4333 vcpu->mmio_fragments[0].gpa, *(u64 *)val); read_prepare() 4341 static int read_emulate(struct kvm_vcpu *vcpu, gpa_t gpa, read_emulate() argument 4344 return !kvm_vcpu_read_guest(vcpu, gpa, val, bytes); read_emulate() 4347 static int write_emulate(struct kvm_vcpu *vcpu, gpa_t gpa, write_emulate() argument 4350 return emulator_write_phys(vcpu, gpa, val, bytes); write_emulate() 4353 static int write_mmio(struct kvm_vcpu *vcpu, gpa_t gpa, int bytes, void *val) write_mmio() argument 4355 trace_kvm_mmio(KVM_TRACE_MMIO_WRITE, bytes, gpa, *(u64 *)val); write_mmio() 4356 return vcpu_mmio_write(vcpu, gpa, bytes, val); write_mmio() 4359 static int read_exit_mmio(struct kvm_vcpu *vcpu, gpa_t gpa, read_exit_mmio() argument 4362 trace_kvm_mmio(KVM_TRACE_MMIO_READ_UNSATISFIED, bytes, gpa, 0); read_exit_mmio() 4366 static int write_exit_mmio(struct kvm_vcpu *vcpu, gpa_t gpa, write_exit_mmio() argument 4395 gpa_t gpa; emulator_read_write_onepage() local 4400 ret = vcpu_mmio_gva_to_gpa(vcpu, addr, &gpa, exception, write); emulator_read_write_onepage() 4409 if (ops->read_write_emulate(vcpu, gpa, val, bytes)) emulator_read_write_onepage() 4416 handled = ops->read_write_mmio(vcpu, gpa, bytes, val); emulator_read_write_onepage() 4420 gpa += handled; emulator_read_write_onepage() 4426 frag->gpa = gpa; emulator_read_write_onepage() 4439 gpa_t gpa; emulator_read_write() local 4473 gpa = vcpu->mmio_fragments[0].gpa; emulator_read_write() 4481 vcpu->run->mmio.phys_addr = gpa; emulator_read_write() 4483 return ops->read_write_exit_mmio(vcpu, gpa, val, bytes); emulator_read_write() 4524 gpa_t gpa; emulator_cmpxchg_emulated() local 4533 gpa = kvm_mmu_gva_to_gpa_write(vcpu, addr, NULL); emulator_cmpxchg_emulated() 4535 if (gpa == UNMAPPED_GVA || emulator_cmpxchg_emulated() 4536 (gpa & PAGE_MASK) == APIC_DEFAULT_PHYS_BASE) emulator_cmpxchg_emulated() 4539 if (((gpa + bytes - 1) & PAGE_MASK) != (gpa & PAGE_MASK)) emulator_cmpxchg_emulated() 4542 page = kvm_vcpu_gfn_to_page(vcpu, gpa >> PAGE_SHIFT); emulator_cmpxchg_emulated() 4547 kaddr += offset_in_page(gpa); emulator_cmpxchg_emulated() 4570 kvm_vcpu_mark_page_dirty(vcpu, gpa >> PAGE_SHIFT); emulator_cmpxchg_emulated() 4571 kvm_mmu_pte_write(vcpu, gpa, new, bytes); emulator_cmpxchg_emulated() 5114 gpa_t gpa = cr2; reexecute_instruction() local 5125 gpa = kvm_mmu_gva_to_gpa_write(vcpu, cr2, NULL); reexecute_instruction() 5131 if (gpa == UNMAPPED_GVA) reexecute_instruction() 5141 pfn = gfn_to_pfn(vcpu->kvm, gpa_to_gfn(gpa)); reexecute_instruction() 5161 kvm_mmu_unprotect_page(vcpu->kvm, gpa_to_gfn(gpa)); reexecute_instruction() 5171 kvm_mmu_unprotect_page(vcpu->kvm, gpa_to_gfn(gpa)); reexecute_instruction() 5185 unsigned long last_retry_eip, last_retry_addr, gpa = cr2; retry_instruction() local 5218 gpa = kvm_mmu_gva_to_gpa_write(vcpu, cr2, NULL); retry_instruction() 5220 kvm_mmu_unprotect_page(vcpu->kvm, gpa_to_gfn(gpa)); retry_instruction() 6733 * write gpa, len 6741 * write gpa, len 6766 frag->gpa += len; complete_emulated_mmio() 6781 run->mmio.phys_addr = frag->gpa; complete_emulated_mmio() 7151 gpa_t gpa; kvm_arch_vcpu_ioctl_translate() local 7155 gpa = kvm_mmu_gva_to_gpa_system(vcpu, vaddr, NULL); kvm_arch_vcpu_ioctl_translate() 7157 tr->physical_address = gpa; kvm_arch_vcpu_ioctl_translate() 7158 tr->valid = gpa != UNMAPPED_GVA; kvm_arch_vcpu_ioctl_translate() 7696 int __x86_set_memory_region(struct kvm *kvm, int id, gpa_t gpa, u32 size) __x86_set_memory_region() argument 7733 m.guest_phys_addr = gpa; __x86_set_memory_region() 7750 int x86_set_memory_region(struct kvm *kvm, int id, gpa_t gpa, u32 size) x86_set_memory_region() argument 7755 r = __x86_set_memory_region(kvm, id, gpa, size); x86_set_memory_region() 4272 vcpu_mmio_gva_to_gpa(struct kvm_vcpu *vcpu, unsigned long gva, gpa_t *gpa, struct x86_exception *exception, bool write) vcpu_mmio_gva_to_gpa() argument
|
H A D | paging_tmpl.h | 878 gpa_t gpa = UNMAPPED_GVA; gva_to_gpa() local 884 gpa = gfn_to_gpa(walker.gfn); gva_to_gpa() 885 gpa |= vaddr & ~PAGE_MASK; gva_to_gpa() 889 return gpa; gva_to_gpa() 898 gpa_t gpa = UNMAPPED_GVA; gva_to_gpa_nested() local 904 gpa = gfn_to_gpa(walker.gfn); gva_to_gpa_nested() 905 gpa |= vaddr & ~PAGE_MASK; gva_to_gpa_nested() 909 return gpa; gva_to_gpa_nested()
|
H A D | mmu.c | 3481 static int tdp_page_fault(struct kvm_vcpu *vcpu, gva_t gpa, u32 error_code, tdp_page_fault() argument 3488 gfn_t gfn = gpa >> PAGE_SHIFT; tdp_page_fault() 3496 r = handle_mmio_page_fault(vcpu, gpa, true); tdp_page_fault() 3516 if (fast_page_fault(vcpu, gpa, level, error_code)) tdp_page_fault() 3522 if (try_async_pf(vcpu, prefault, gfn, gpa, &pfn, write, &map_writable)) tdp_page_fault() 3534 r = __direct_map(vcpu, gpa, write, map_writable, tdp_page_fault() 4173 static u64 mmu_pte_write_fetch_gpte(struct kvm_vcpu *vcpu, gpa_t *gpa, mmu_pte_write_fetch_gpte() argument 4186 *gpa &= ~(gpa_t)7; mmu_pte_write_fetch_gpte() 4188 r = kvm_vcpu_read_guest(vcpu, *gpa, &gentry, 8); mmu_pte_write_fetch_gpte() 4229 static bool detect_write_misaligned(struct kvm_mmu_page *sp, gpa_t gpa, detect_write_misaligned() argument 4234 pgprintk("misaligned: gpa %llx bytes %d role %x\n", detect_write_misaligned() 4235 gpa, bytes, sp->role.word); detect_write_misaligned() 4237 offset = offset_in_page(gpa); detect_write_misaligned() 4253 static u64 *get_written_sptes(struct kvm_mmu_page *sp, gpa_t gpa, int *nspte) get_written_sptes() argument 4259 page_offset = offset_in_page(gpa); get_written_sptes() 4284 void kvm_mmu_pte_write(struct kvm_vcpu *vcpu, gpa_t gpa, kvm_mmu_pte_write() argument 4287 gfn_t gfn = gpa >> PAGE_SHIFT; kvm_mmu_pte_write() 4311 pgprintk("%s: gpa %llx bytes %d\n", __func__, gpa, bytes); kvm_mmu_pte_write() 4313 gentry = mmu_pte_write_fetch_gpte(vcpu, &gpa, new, &bytes); kvm_mmu_pte_write() 4327 if (detect_write_misaligned(sp, gpa, bytes) || kvm_mmu_pte_write() 4335 spte = get_written_sptes(sp, gpa, &npte); kvm_mmu_pte_write() 4360 gpa_t gpa; kvm_mmu_unprotect_page_virt() local 4366 gpa = kvm_mmu_gva_to_gpa_read(vcpu, gva, NULL); kvm_mmu_unprotect_page_virt() 4368 r = kvm_mmu_unprotect_page(vcpu->kvm, gpa >> PAGE_SHIFT); kvm_mmu_unprotect_page_virt()
|
H A D | vmx.c | 1286 static inline void __invept(int ext, u64 eptp, gpa_t gpa) __invept() argument 1289 u64 eptp, gpa; __invept() member in struct:__anon3202 1290 } operand = {eptp, gpa}; __invept() 5880 gpa_t gpa; handle_ept_violation() local 5910 gpa = vmcs_read64(GUEST_PHYSICAL_ADDRESS); handle_ept_violation() 5911 trace_kvm_page_fault(gpa, exit_qualification); handle_ept_violation() 5922 return kvm_mmu_page_fault(vcpu, gpa, error_code, NULL, 0); handle_ept_violation() 5928 gpa_t gpa; handle_ept_misconfig() local 5930 gpa = vmcs_read64(GUEST_PHYSICAL_ADDRESS); handle_ept_misconfig() 5931 if (!kvm_io_bus_write(vcpu, KVM_FAST_MMIO_BUS, gpa, 0, NULL)) { handle_ept_misconfig() 5933 trace_kvm_fast_mmio(gpa); handle_ept_misconfig() 5937 ret = handle_mmio_page_fault(vcpu, gpa, true); handle_ept_misconfig() 5939 return x86_emulate_instruction(vcpu, gpa, 0, NULL, 0) == handle_ept_misconfig() 5943 return kvm_mmu_page_fault(vcpu, gpa, 0, NULL, 0); handle_ept_misconfig() 7324 u64 eptp, gpa; handle_invept() member in struct:__anon3203 7886 u64 gpa; vmx_flush_pml_buffer() local 7888 gpa = pml_buf[pml_idx]; vmx_flush_pml_buffer() 7889 WARN_ON(gpa & (PAGE_SIZE - 1)); vmx_flush_pml_buffer() 7890 kvm_vcpu_mark_page_dirty(vcpu, gpa >> PAGE_SHIFT); vmx_flush_pml_buffer() 9382 static u32 nested_vmx_load_msr(struct kvm_vcpu *vcpu, u64 gpa, u32 count) nested_vmx_load_msr() argument 9390 if (kvm_vcpu_read_guest(vcpu, gpa + i * sizeof(e), nested_vmx_load_msr() 9394 __func__, i, gpa + i * sizeof(e)); nested_vmx_load_msr() 9417 static int nested_vmx_store_msr(struct kvm_vcpu *vcpu, u64 gpa, u32 count) nested_vmx_store_msr() argument 9425 gpa + i * sizeof(e), nested_vmx_store_msr() 9429 __func__, i, gpa + i * sizeof(e)); nested_vmx_store_msr() 9447 gpa + i * sizeof(e) + nested_vmx_store_msr()
|
H A D | svm.c | 104 /* gpa pointers to the real vectors */ 2028 static void *nested_svm_map(struct vcpu_svm *svm, u64 gpa, struct page **_page) nested_svm_map() argument 2034 page = kvm_vcpu_gfn_to_page(&svm->vcpu, gpa >> PAGE_SHIFT); nested_svm_map() 2059 u64 gpa; nested_svm_intercept_ioio() local 2067 gpa = svm->nested.vmcb_iopm + (port / 8); nested_svm_intercept_ioio() 2073 if (kvm_vcpu_read_guest(&svm->vcpu, gpa, &val, iopm_len)) nested_svm_intercept_ioio()
|
/linux-4.4.14/drivers/misc/sgi-xp/ |
H A D | xp_uv.c | 39 xp_socket_pa_uv(unsigned long gpa) xp_socket_pa_uv() argument 41 return uv_gpa_to_soc_phys_ram(gpa); xp_socket_pa_uv()
|
H A D | xp_sn2.c | 90 xp_socket_pa_sn2(unsigned long gpa) xp_socket_pa_sn2() argument 92 return gpa; xp_socket_pa_sn2()
|
H A D | xp_main.c | 47 unsigned long (*xp_socket_pa) (unsigned long gpa);
|
/linux-4.4.14/virt/kvm/ |
H A D | kvm_main.c | 1656 int kvm_read_guest(struct kvm *kvm, gpa_t gpa, void *data, unsigned long len) kvm_read_guest() argument 1658 gfn_t gfn = gpa >> PAGE_SHIFT; kvm_read_guest() 1660 int offset = offset_in_page(gpa); kvm_read_guest() 1676 int kvm_vcpu_read_guest(struct kvm_vcpu *vcpu, gpa_t gpa, void *data, unsigned long len) kvm_vcpu_read_guest() argument 1678 gfn_t gfn = gpa >> PAGE_SHIFT; kvm_vcpu_read_guest() 1680 int offset = offset_in_page(gpa); kvm_vcpu_read_guest() 1713 int kvm_read_guest_atomic(struct kvm *kvm, gpa_t gpa, void *data, kvm_read_guest_atomic() argument 1716 gfn_t gfn = gpa >> PAGE_SHIFT; kvm_read_guest_atomic() 1718 int offset = offset_in_page(gpa); kvm_read_guest_atomic() 1724 int kvm_vcpu_read_guest_atomic(struct kvm_vcpu *vcpu, gpa_t gpa, kvm_vcpu_read_guest_atomic() argument 1727 gfn_t gfn = gpa >> PAGE_SHIFT; kvm_vcpu_read_guest_atomic() 1729 int offset = offset_in_page(gpa); kvm_vcpu_read_guest_atomic() 1769 int kvm_write_guest(struct kvm *kvm, gpa_t gpa, const void *data, kvm_write_guest() argument 1772 gfn_t gfn = gpa >> PAGE_SHIFT; kvm_write_guest() 1774 int offset = offset_in_page(gpa); kvm_write_guest() 1790 int kvm_vcpu_write_guest(struct kvm_vcpu *vcpu, gpa_t gpa, const void *data, kvm_vcpu_write_guest() argument 1793 gfn_t gfn = gpa >> PAGE_SHIFT; kvm_vcpu_write_guest() 1795 int offset = offset_in_page(gpa); kvm_vcpu_write_guest() 1812 gpa_t gpa, unsigned long len) kvm_gfn_to_hva_cache_init() 1815 int offset = offset_in_page(gpa); kvm_gfn_to_hva_cache_init() 1816 gfn_t start_gfn = gpa >> PAGE_SHIFT; kvm_gfn_to_hva_cache_init() 1817 gfn_t end_gfn = (gpa + len - 1) >> PAGE_SHIFT; kvm_gfn_to_hva_cache_init() 1821 ghc->gpa = gpa; kvm_gfn_to_hva_cache_init() 1857 kvm_gfn_to_hva_cache_init(kvm, ghc, ghc->gpa, ghc->len); kvm_write_guest_cached() 1860 return kvm_write_guest(kvm, ghc->gpa, data, len); kvm_write_guest_cached() 1868 mark_page_dirty_in_slot(ghc->memslot, ghc->gpa >> PAGE_SHIFT); kvm_write_guest_cached() 1883 kvm_gfn_to_hva_cache_init(kvm, ghc, ghc->gpa, ghc->len); kvm_read_guest_cached() 1886 return kvm_read_guest(kvm, ghc->gpa, data, len); kvm_read_guest_cached() 1907 int kvm_clear_guest(struct kvm *kvm, gpa_t gpa, unsigned long len) kvm_clear_guest() argument 1909 gfn_t gfn = gpa >> PAGE_SHIFT; kvm_clear_guest() 1911 int offset = offset_in_page(gpa); kvm_clear_guest() 1811 kvm_gfn_to_hva_cache_init(struct kvm *kvm, struct gfn_to_hva_cache *ghc, gpa_t gpa, unsigned long len) kvm_gfn_to_hva_cache_init() argument
|
/linux-4.4.14/include/xen/interface/hvm/ |
H A D | hvm_op.h | 41 aligned_u64 gpa; member in struct:xen_hvm_pagetable_dying
|
/linux-4.4.14/drivers/misc/sgi-gru/ |
H A D | grufault.c | 262 int write, int atomic, unsigned long *gpa, int *pageshift) gru_vtop() 288 *gpa = uv_soc_phys_ram_to_gpa(paddr); gru_vtop() 325 unsigned long vaddr = 0, gpa; gru_preload_tlb() local 341 ret = gru_vtop(gts, vaddr, write, atomic, &gpa, &pageshift); gru_preload_tlb() 342 if (ret || tfh_write_only(tfh, gpa, GAA_RAM, vaddr, asid, write, gru_preload_tlb() 346 "%s: gid %d, gts 0x%p, tfh 0x%p, vaddr 0x%lx, asid 0x%x, rw %d, ps %d, gpa 0x%lx\n", gru_preload_tlb() 348 vaddr, asid, write, pageshift, gpa); gru_preload_tlb() 372 unsigned long gpa = 0, vaddr = 0; gru_try_dropin() local 422 ret = gru_vtop(gts, vaddr, write, atomic, &gpa, &pageshift); gru_try_dropin() 443 tfh_write_restart(tfh, gpa, GAA_RAM, vaddr, asid, write, gru_try_dropin() 447 " rw %d, ps %d, gpa 0x%lx\n", gru_try_dropin() 449 indexway, write, pageshift, gpa); gru_try_dropin() 261 gru_vtop(struct gru_thread_state *gts, unsigned long vaddr, int write, int atomic, unsigned long *gpa, int *pageshift) gru_vtop() argument
|
H A D | grukservices.h | 138 * gpa source numalink physical address of GPA 144 int gru_read_gpa(unsigned long *value, unsigned long gpa);
|
H A D | gru_instructions.h | 359 static inline void gru_vload_phys(void *cb, unsigned long gpa, gru_vload_phys() argument 364 ins->baddr0 = (long)gpa | ((unsigned long)iaa << 62); gru_vload_phys() 371 static inline void gru_vstore_phys(void *cb, unsigned long gpa, gru_vstore_phys() argument 376 ins->baddr0 = (long)gpa | ((unsigned long)iaa << 62); gru_vstore_phys()
|
H A D | grukservices.c | 721 unsigned long m, *val = mesg, gpa, save; send_message_put_nacked() local 744 gpa = uv_global_gru_mmr_address(mqd->interrupt_pnode, UVH_IPI_INT); send_message_put_nacked() 748 gru_vstore_phys(cb, gpa, gru_get_tri(mesg), IAA_REGISTER, IMA); send_message_put_nacked() 911 int gru_read_gpa(unsigned long *value, unsigned long gpa) gru_read_gpa() argument 920 iaa = gpa >> 62; gru_read_gpa() 921 gru_vload_phys(cb, gpa, gru_get_tri(dsr), iaa, IMA); gru_read_gpa()
|
/linux-4.4.14/arch/powerpc/kvm/ |
H A D | e500.h | 258 gpa_t gpa; tlbe_is_host_safe() local 270 gpa = get_tlb_raddr(tlbe); tlbe_is_host_safe() 271 if (!gfn_to_memslot(vcpu->kvm, gpa >> PAGE_SHIFT)) tlbe_is_host_safe()
|
H A D | book3s_64_mmu_hv.c | 392 unsigned long gpa, gva_t ea, int is_store) kvmppc_hv_emulate_mmio() 431 vcpu->arch.paddr_accessed = gpa; kvmppc_hv_emulate_mmio() 444 unsigned long gpa, gfn, hva, pfn; kvmppc_book3s_hv_page_fault() local 483 gpa = gpa_base | (ea & (psize - 1)); kvmppc_book3s_hv_page_fault() 484 gfn = gpa >> PAGE_SHIFT; kvmppc_book3s_hv_page_fault() 491 return kvmppc_hv_emulate_mmio(run, vcpu, gpa, ea, kvmppc_book3s_hv_page_fault() 1018 gfn = vpa->gpa >> PAGE_SHIFT; harvest_vpa_dirty() 1062 void *kvmppc_pin_guest_page(struct kvm *kvm, unsigned long gpa, kvmppc_pin_guest_page() argument 1066 unsigned long gfn = gpa >> PAGE_SHIFT; kvmppc_pin_guest_page() 1083 offset = gpa & (PAGE_SIZE - 1); kvmppc_pin_guest_page() 1093 void kvmppc_unpin_guest_page(struct kvm *kvm, void *va, unsigned long gpa, kvmppc_unpin_guest_page() argument 1108 gfn = gpa >> PAGE_SHIFT; kvmppc_unpin_guest_page() 391 kvmppc_hv_emulate_mmio(struct kvm_run *run, struct kvm_vcpu *vcpu, unsigned long gpa, gva_t ea, int is_store) kvmppc_hv_emulate_mmio() argument
|
H A D | book3s.c | 369 pfn_t kvmppc_gpa_to_pfn(struct kvm_vcpu *vcpu, gpa_t gpa, bool writing, kvmppc_gpa_to_pfn() argument 373 gfn_t gfn = gpa >> PAGE_SHIFT; kvmppc_gpa_to_pfn() 379 gpa &= ~0xFFFULL; kvmppc_gpa_to_pfn() 380 if (unlikely(mp_pa) && unlikely((gpa & KVM_PAM) == mp_pa)) { kvmppc_gpa_to_pfn()
|
H A D | book3s_32_mmu_host.c | 159 /* Get host physical address for gpa */ kvmppc_mmu_map_page() 162 printk(KERN_INFO "Couldn't get guest page for gpa %lx!\n", kvmppc_mmu_map_page()
|
H A D | book3s_64_mmu_host.c | 107 /* Get host physical address for gpa */ kvmppc_mmu_map_page() 110 printk(KERN_INFO "Couldn't get guest page for gpa %lx!\n", kvmppc_mmu_map_page()
|
H A D | book3s_hv.c | 484 unsigned long gpa; kvmppc_update_vpa() local 495 gpa = vpap->next_gpa; kvmppc_update_vpa() 499 if (gpa) kvmppc_update_vpa() 500 va = kvmppc_pin_guest_page(kvm, gpa, &nb); kvmppc_update_vpa() 502 if (gpa == vpap->next_gpa) kvmppc_update_vpa() 506 kvmppc_unpin_guest_page(kvm, va, gpa, false); kvmppc_update_vpa() 516 kvmppc_unpin_guest_page(kvm, va, gpa, false); kvmppc_update_vpa() 520 kvmppc_unpin_guest_page(kvm, vpap->pinned_addr, vpap->gpa, kvmppc_update_vpa() 522 vpap->gpa = gpa; kvmppc_update_vpa() 1691 kvmppc_unpin_guest_page(kvm, vpa->pinned_addr, vpa->gpa, unpin_vpa()
|
H A D | book3s_hv_rm_mmu.c | 171 unsigned long i, pa, gpa, gfn, psize; kvmppc_do_h_enter() local 198 gpa = (ptel & HPTE_R_RPN) & ~(psize - 1); kvmppc_do_h_enter() 199 gfn = gpa >> PAGE_SHIFT; kvmppc_do_h_enter() 256 pa |= gpa & ~PAGE_MASK; kvmppc_do_h_enter()
|
H A D | book3s_pr.c | 515 static int kvmppc_visible_gpa(struct kvm_vcpu *vcpu, gpa_t gpa) kvmppc_visible_gpa() argument 522 gpa &= ~0xFFFULL; kvmppc_visible_gpa() 523 if (unlikely(mp_pa) && unlikely((mp_pa & KVM_PAM) == (gpa & KVM_PAM))) { kvmppc_visible_gpa() 527 return kvm_is_visible_gfn(vcpu->kvm, gpa >> PAGE_SHIFT); kvmppc_visible_gpa()
|
/linux-4.4.14/arch/arm/kvm/ |
H A D | mmu.c | 778 gpa_t gpa = addr + (vm_start - memslot->userspace_addr); stage2_unmap_memslot() local 779 unmap_stage2_range(kvm, gpa, vm_end - vm_start); stage2_unmap_memslot() 1470 gpa_t gpa, void *data), handle_hva_to_gpa() 1498 gpa_t gpa = gfn << PAGE_SHIFT; kvm_for_each_memslot() local 1499 ret |= handler(kvm, gpa, data); kvm_for_each_memslot() 1506 static int kvm_unmap_hva_handler(struct kvm *kvm, gpa_t gpa, void *data) kvm_unmap_hva_handler() argument 1508 unmap_stage2_range(kvm, gpa, PAGE_SIZE); kvm_unmap_hva_handler() 1535 static int kvm_set_spte_handler(struct kvm *kvm, gpa_t gpa, void *data) kvm_set_spte_handler() argument 1546 stage2_set_pte(kvm, NULL, gpa, pte, 0); kvm_set_spte_handler() 1564 static int kvm_age_hva_handler(struct kvm *kvm, gpa_t gpa, void *data) kvm_age_hva_handler() argument 1569 pmd = stage2_get_pmd(kvm, NULL, gpa); kvm_age_hva_handler() 1582 pte = pte_offset_kernel(pmd, gpa); kvm_age_hva_handler() 1594 static int kvm_test_age_hva_handler(struct kvm *kvm, gpa_t gpa, void *data) kvm_test_age_hva_handler() argument 1599 pmd = stage2_get_pmd(kvm, NULL, gpa); kvm_test_age_hva_handler() 1606 pte = pte_offset_kernel(pmd, gpa); kvm_test_age_hva_handler() 1797 gpa_t gpa = mem->guest_phys_addr + kvm_arch_prepare_memory_region() local 1808 ret = kvm_phys_addr_ioremap(kvm, gpa, pa, kvm_arch_prepare_memory_region() 1860 gpa_t gpa = slot->base_gfn << PAGE_SHIFT; kvm_arch_flush_shadow_memslot() local 1864 unmap_stage2_range(kvm, gpa, size); kvm_arch_flush_shadow_memslot() 1466 handle_hva_to_gpa(struct kvm *kvm, unsigned long start, unsigned long end, int (*handler)(struct kvm *kvm, gpa_t gpa, void *data), void *data) handle_hva_to_gpa() argument
|
/linux-4.4.14/arch/mips/kvm/ |
H A D | trap_emul.c | 24 gpa_t gpa; kvm_trap_emul_gva_to_gpa_cb() local 28 gpa = CPHYSADDR(gva); kvm_trap_emul_gva_to_gpa_cb() 32 gpa = KVM_INVALID_ADDR; kvm_trap_emul_gva_to_gpa_cb() 35 kvm_debug("%s: gva %#lx, gpa: %#llx\n", __func__, gva, gpa); kvm_trap_emul_gva_to_gpa_cb() 37 return gpa; kvm_trap_emul_gva_to_gpa_cb()
|
/linux-4.4.14/drivers/pinctrl/samsung/ |
H A D | pinctrl-s3c24xx.c | 562 PIN_BANK_A(23, 0x000, "gpa"), 582 PIN_BANK_A(27, 0x000, "gpa"), 604 PIN_BANK_A(25, 0x000, "gpa"), 624 PIN_BANK_A(28, 0x000, "gpa"),
|
H A D | pinctrl-s3c64xx.c | 780 PIN_BANK_4BIT_EINTG(8, 0x000, "gpa", 0),
|
/linux-4.4.14/tools/perf/arch/x86/util/ |
H A D | kvm-stat.c | 22 key->key = perf_evsel__intval(evsel, sample, "gpa"); mmio_event_get_key()
|
/linux-4.4.14/arch/x86/include/asm/ |
H A D | kvm_host.h | 279 gpa_t (*translate_gpa)(struct kvm_vcpu *vcpu, gpa_t gpa, u32 access, 955 int emulator_write_phys(struct kvm_vcpu *vcpu, gpa_t gpa, 1070 void kvm_mmu_pte_write(struct kvm_vcpu *vcpu, gpa_t gpa, 1078 gpa_t translate_nested_gpa(struct kvm_vcpu *vcpu, gpa_t gpa, u32 access, 1099 static inline gpa_t translate_gpa(struct kvm_vcpu *vcpu, gpa_t gpa, u32 access, translate_gpa() argument 1102 return gpa; translate_gpa() 1247 int __x86_set_memory_region(struct kvm *kvm, int id, gpa_t gpa, u32 size); 1248 int x86_set_memory_region(struct kvm *kvm, int id, gpa_t gpa, u32 size);
|
H A D | kvm_emulate.h | 25 u64 address; /* cr2 or nested page fault gpa */
|
/linux-4.4.14/arch/ia64/include/asm/uv/ |
H A D | uv_hub.h | 166 static inline void *uv_va(unsigned long gpa) uv_va() argument 168 return __va(gpa & uv_hub_info->gpa_mask); uv_va()
|
/linux-4.4.14/arch/powerpc/include/asm/ |
H A D | kvm_book3s.h | 157 extern pfn_t kvmppc_gpa_to_pfn(struct kvm_vcpu *vcpu, gpa_t gpa, bool writing, 169 unsigned long gpa, bool dirty);
|
H A D | kvm_host.h | 329 unsigned long gpa; /* Current guest phys addr */ member in struct:kvmppc_vpa
|
/linux-4.4.14/arch/x86/platform/uv/ |
H A D | tlb_uv.c | 1693 unsigned long gpa; activation_descriptor_init() local 1711 gpa = uv_gpa(bau_desc); activation_descriptor_init() 1712 n = uv_gpa_to_gnode(gpa); activation_descriptor_init() 1713 m = uv_gpa_to_offset(gpa); activation_descriptor_init()
|
/linux-4.4.14/kernel/rcu/ |
H A D | tree.c | 1194 unsigned long gpa; rcu_check_gp_kthread_starvation() local 1198 gpa = READ_ONCE(rsp->gp_activity); rcu_check_gp_kthread_starvation() 1199 if (j - gpa > 2 * HZ) rcu_check_gp_kthread_starvation() 1201 rsp->name, j - gpa, rcu_check_gp_kthread_starvation() 1232 unsigned long gpa; print_other_cpu_stall() local 1286 gpa = READ_ONCE(rsp->gp_activity); 1288 rsp->name, j - gpa, j, gpa,
|
/linux-4.4.14/arch/x86/xen/ |
H A D | mmu.c | 2751 a.gpa = __pa(mm->pgd); xen_hvm_exit_mmap() 2762 a.gpa = 0x00; is_pagetable_dying_supported()
|