Lines Matching refs:gpa
778 gpa_t gpa = addr + (vm_start - memslot->userspace_addr); in stage2_unmap_memslot() local
779 unmap_stage2_range(kvm, gpa, vm_end - vm_start); in stage2_unmap_memslot()
1470 gpa_t gpa, void *data), in handle_hva_to_gpa() argument
1498 gpa_t gpa = gfn << PAGE_SHIFT; in handle_hva_to_gpa() local
1499 ret |= handler(kvm, gpa, data); in handle_hva_to_gpa()
1506 static int kvm_unmap_hva_handler(struct kvm *kvm, gpa_t gpa, void *data) in kvm_unmap_hva_handler() argument
1508 unmap_stage2_range(kvm, gpa, PAGE_SIZE); in kvm_unmap_hva_handler()
1535 static int kvm_set_spte_handler(struct kvm *kvm, gpa_t gpa, void *data) in kvm_set_spte_handler() argument
1546 stage2_set_pte(kvm, NULL, gpa, pte, 0); in kvm_set_spte_handler()
1564 static int kvm_age_hva_handler(struct kvm *kvm, gpa_t gpa, void *data) in kvm_age_hva_handler() argument
1569 pmd = stage2_get_pmd(kvm, NULL, gpa); in kvm_age_hva_handler()
1582 pte = pte_offset_kernel(pmd, gpa); in kvm_age_hva_handler()
1594 static int kvm_test_age_hva_handler(struct kvm *kvm, gpa_t gpa, void *data) in kvm_test_age_hva_handler() argument
1599 pmd = stage2_get_pmd(kvm, NULL, gpa); in kvm_test_age_hva_handler()
1606 pte = pte_offset_kernel(pmd, gpa); in kvm_test_age_hva_handler()
1797 gpa_t gpa = mem->guest_phys_addr + in kvm_arch_prepare_memory_region() local
1808 ret = kvm_phys_addr_ioremap(kvm, gpa, pa, in kvm_arch_prepare_memory_region()
1860 gpa_t gpa = slot->base_gfn << PAGE_SHIFT; in kvm_arch_flush_shadow_memslot() local
1864 unmap_stage2_range(kvm, gpa, size); in kvm_arch_flush_shadow_memslot()