host_ctxt 144 arch/arm/include/asm/kvm_host.h struct kvm_cpu_context host_ctxt; host_ctxt 148 arch/arm/kvm/hyp/switch.c struct kvm_cpu_context *host_ctxt; host_ctxt 157 arch/arm/kvm/hyp/switch.c host_ctxt = kern_hyp_va(vcpu->arch.host_cpu_context); host_ctxt 160 arch/arm/kvm/hyp/switch.c __sysreg_save_state(host_ctxt); host_ctxt 161 arch/arm/kvm/hyp/switch.c __banked_save_state(host_ctxt); host_ctxt 174 arch/arm/kvm/hyp/switch.c exit_code = __guest_enter(vcpu, host_ctxt); host_ctxt 191 arch/arm/kvm/hyp/switch.c __banked_restore_state(host_ctxt); host_ctxt 192 arch/arm/kvm/hyp/switch.c __sysreg_restore_state(host_ctxt); host_ctxt 196 arch/arm/kvm/hyp/switch.c __vfp_restore_state(&host_ctxt->vfp); host_ctxt 227 arch/arm/kvm/hyp/switch.c struct kvm_cpu_context *host_ctxt; host_ctxt 230 arch/arm/kvm/hyp/switch.c host_ctxt = kern_hyp_va(vcpu->arch.host_cpu_context); host_ctxt 234 arch/arm/kvm/hyp/switch.c __banked_restore_state(host_ctxt); host_ctxt 235 arch/arm/kvm/hyp/switch.c __sysreg_restore_state(host_ctxt); host_ctxt 231 arch/arm64/include/asm/kvm_host.h struct kvm_cpu_context host_ctxt; host_ctxt 81 arch/arm64/include/asm/kvm_hyp.h u64 __guest_enter(struct kvm_vcpu *vcpu, struct kvm_cpu_context *host_ctxt); host_ctxt 104 arch/arm64/kernel/asm-offsets.c DEFINE(HOST_DATA_CONTEXT, offsetof(struct kvm_host_data, host_ctxt)); host_ctxt 173 arch/arm64/kvm/hyp/debug-sr.c struct kvm_cpu_context *host_ctxt; host_ctxt 188 arch/arm64/kvm/hyp/debug-sr.c host_ctxt = kern_hyp_va(vcpu->arch.host_cpu_context); host_ctxt 193 arch/arm64/kvm/hyp/debug-sr.c __debug_save_state(vcpu, host_dbg, host_ctxt); host_ctxt 199 arch/arm64/kvm/hyp/debug-sr.c struct kvm_cpu_context *host_ctxt; host_ctxt 210 arch/arm64/kvm/hyp/debug-sr.c host_ctxt = kern_hyp_va(vcpu->arch.host_cpu_context); host_ctxt 216 arch/arm64/kvm/hyp/debug-sr.c __debug_restore_state(vcpu, host_dbg, host_ctxt); host_ctxt 561 arch/arm64/kvm/hyp/switch.c static bool __hyp_text __pmu_switch_to_guest(struct kvm_cpu_context *host_ctxt) host_ctxt 566 arch/arm64/kvm/hyp/switch.c host = container_of(host_ctxt, struct kvm_host_data, host_ctxt); host_ctxt 581 arch/arm64/kvm/hyp/switch.c static void __hyp_text __pmu_switch_to_host(struct kvm_cpu_context *host_ctxt) host_ctxt 586 arch/arm64/kvm/hyp/switch.c host = container_of(host_ctxt, struct kvm_host_data, host_ctxt); host_ctxt 599 arch/arm64/kvm/hyp/switch.c struct kvm_cpu_context *host_ctxt; host_ctxt 603 arch/arm64/kvm/hyp/switch.c host_ctxt = vcpu->arch.host_cpu_context; host_ctxt 604 arch/arm64/kvm/hyp/switch.c host_ctxt->__hyp_running_vcpu = vcpu; host_ctxt 607 arch/arm64/kvm/hyp/switch.c sysreg_save_host_state_vhe(host_ctxt); host_ctxt 630 arch/arm64/kvm/hyp/switch.c exit_code = __guest_enter(vcpu, host_ctxt); host_ctxt 641 arch/arm64/kvm/hyp/switch.c sysreg_restore_host_state_vhe(host_ctxt); host_ctxt 655 arch/arm64/kvm/hyp/switch.c struct kvm_cpu_context *host_ctxt; host_ctxt 673 arch/arm64/kvm/hyp/switch.c host_ctxt = kern_hyp_va(vcpu->arch.host_cpu_context); host_ctxt 674 arch/arm64/kvm/hyp/switch.c host_ctxt->__hyp_running_vcpu = vcpu; host_ctxt 677 arch/arm64/kvm/hyp/switch.c pmu_switch_needed = __pmu_switch_to_guest(host_ctxt); host_ctxt 679 arch/arm64/kvm/hyp/switch.c __sysreg_save_state_nvhe(host_ctxt); host_ctxt 699 arch/arm64/kvm/hyp/switch.c exit_code = __guest_enter(vcpu, host_ctxt); host_ctxt 714 arch/arm64/kvm/hyp/switch.c __sysreg_restore_state_nvhe(host_ctxt); host_ctxt 726 arch/arm64/kvm/hyp/switch.c __pmu_switch_to_host(host_ctxt); host_ctxt 766 arch/arm64/kvm/hyp/switch.c struct kvm_cpu_context *host_ctxt) host_ctxt 769 arch/arm64/kvm/hyp/switch.c vcpu = host_ctxt->__hyp_running_vcpu; host_ctxt 772 arch/arm64/kvm/hyp/switch.c sysreg_restore_host_state_vhe(host_ctxt); host_ctxt 781 arch/arm64/kvm/hyp/switch.c void __hyp_text __noreturn hyp_panic(struct kvm_cpu_context *host_ctxt) host_ctxt 788 arch/arm64/kvm/hyp/switch.c __hyp_call_panic_nvhe(spsr, elr, par, host_ctxt); host_ctxt 790 arch/arm64/kvm/hyp/switch.c __hyp_call_panic_vhe(spsr, elr, par, host_ctxt); host_ctxt 248 arch/arm64/kvm/hyp/sysreg-sr.c struct kvm_cpu_context *host_ctxt = vcpu->arch.host_cpu_context; host_ctxt 254 arch/arm64/kvm/hyp/sysreg-sr.c __sysreg_save_user_state(host_ctxt); host_ctxt 284 arch/arm64/kvm/hyp/sysreg-sr.c struct kvm_cpu_context *host_ctxt = vcpu->arch.host_cpu_context; host_ctxt 297 arch/arm64/kvm/hyp/sysreg-sr.c __sysreg_restore_user_state(host_ctxt); host_ctxt 166 arch/arm64/kvm/pmu.c struct kvm_cpu_context *host_ctxt; host_ctxt 173 arch/arm64/kvm/pmu.c host_ctxt = vcpu->arch.host_cpu_context; host_ctxt 174 arch/arm64/kvm/pmu.c host = container_of(host_ctxt, struct kvm_host_data, host_ctxt); host_ctxt 187 arch/arm64/kvm/pmu.c struct kvm_cpu_context *host_ctxt; host_ctxt 194 arch/arm64/kvm/pmu.c host_ctxt = vcpu->arch.host_cpu_context; host_ctxt 195 arch/arm64/kvm/pmu.c host = container_of(host_ctxt, struct kvm_host_data, host_ctxt); host_ctxt 385 virt/kvm/arm/arm.c vcpu->arch.host_cpu_context = &cpu_data->host_ctxt; host_ctxt 1365 virt/kvm/arm/arm.c kvm_init_host_cpu_context(&this_cpu_ptr(&kvm_host_data)->host_ctxt);