host_ctxt         144 arch/arm/include/asm/kvm_host.h 	struct kvm_cpu_context host_ctxt;
host_ctxt         148 arch/arm/kvm/hyp/switch.c 	struct kvm_cpu_context *host_ctxt;
host_ctxt         157 arch/arm/kvm/hyp/switch.c 	host_ctxt = kern_hyp_va(vcpu->arch.host_cpu_context);
host_ctxt         160 arch/arm/kvm/hyp/switch.c 	__sysreg_save_state(host_ctxt);
host_ctxt         161 arch/arm/kvm/hyp/switch.c 	__banked_save_state(host_ctxt);
host_ctxt         174 arch/arm/kvm/hyp/switch.c 	exit_code = __guest_enter(vcpu, host_ctxt);
host_ctxt         191 arch/arm/kvm/hyp/switch.c 	__banked_restore_state(host_ctxt);
host_ctxt         192 arch/arm/kvm/hyp/switch.c 	__sysreg_restore_state(host_ctxt);
host_ctxt         196 arch/arm/kvm/hyp/switch.c 		__vfp_restore_state(&host_ctxt->vfp);
host_ctxt         227 arch/arm/kvm/hyp/switch.c 		struct kvm_cpu_context *host_ctxt;
host_ctxt         230 arch/arm/kvm/hyp/switch.c 		host_ctxt = kern_hyp_va(vcpu->arch.host_cpu_context);
host_ctxt         234 arch/arm/kvm/hyp/switch.c 		__banked_restore_state(host_ctxt);
host_ctxt         235 arch/arm/kvm/hyp/switch.c 		__sysreg_restore_state(host_ctxt);
host_ctxt         231 arch/arm64/include/asm/kvm_host.h 	struct kvm_cpu_context host_ctxt;
host_ctxt          81 arch/arm64/include/asm/kvm_hyp.h u64 __guest_enter(struct kvm_vcpu *vcpu, struct kvm_cpu_context *host_ctxt);
host_ctxt         104 arch/arm64/kernel/asm-offsets.c   DEFINE(HOST_DATA_CONTEXT,	offsetof(struct kvm_host_data, host_ctxt));
host_ctxt         173 arch/arm64/kvm/hyp/debug-sr.c 	struct kvm_cpu_context *host_ctxt;
host_ctxt         188 arch/arm64/kvm/hyp/debug-sr.c 	host_ctxt = kern_hyp_va(vcpu->arch.host_cpu_context);
host_ctxt         193 arch/arm64/kvm/hyp/debug-sr.c 	__debug_save_state(vcpu, host_dbg, host_ctxt);
host_ctxt         199 arch/arm64/kvm/hyp/debug-sr.c 	struct kvm_cpu_context *host_ctxt;
host_ctxt         210 arch/arm64/kvm/hyp/debug-sr.c 	host_ctxt = kern_hyp_va(vcpu->arch.host_cpu_context);
host_ctxt         216 arch/arm64/kvm/hyp/debug-sr.c 	__debug_restore_state(vcpu, host_dbg, host_ctxt);
host_ctxt         561 arch/arm64/kvm/hyp/switch.c static bool __hyp_text __pmu_switch_to_guest(struct kvm_cpu_context *host_ctxt)
host_ctxt         566 arch/arm64/kvm/hyp/switch.c 	host = container_of(host_ctxt, struct kvm_host_data, host_ctxt);
host_ctxt         581 arch/arm64/kvm/hyp/switch.c static void __hyp_text __pmu_switch_to_host(struct kvm_cpu_context *host_ctxt)
host_ctxt         586 arch/arm64/kvm/hyp/switch.c 	host = container_of(host_ctxt, struct kvm_host_data, host_ctxt);
host_ctxt         599 arch/arm64/kvm/hyp/switch.c 	struct kvm_cpu_context *host_ctxt;
host_ctxt         603 arch/arm64/kvm/hyp/switch.c 	host_ctxt = vcpu->arch.host_cpu_context;
host_ctxt         604 arch/arm64/kvm/hyp/switch.c 	host_ctxt->__hyp_running_vcpu = vcpu;
host_ctxt         607 arch/arm64/kvm/hyp/switch.c 	sysreg_save_host_state_vhe(host_ctxt);
host_ctxt         630 arch/arm64/kvm/hyp/switch.c 		exit_code = __guest_enter(vcpu, host_ctxt);
host_ctxt         641 arch/arm64/kvm/hyp/switch.c 	sysreg_restore_host_state_vhe(host_ctxt);
host_ctxt         655 arch/arm64/kvm/hyp/switch.c 	struct kvm_cpu_context *host_ctxt;
host_ctxt         673 arch/arm64/kvm/hyp/switch.c 	host_ctxt = kern_hyp_va(vcpu->arch.host_cpu_context);
host_ctxt         674 arch/arm64/kvm/hyp/switch.c 	host_ctxt->__hyp_running_vcpu = vcpu;
host_ctxt         677 arch/arm64/kvm/hyp/switch.c 	pmu_switch_needed = __pmu_switch_to_guest(host_ctxt);
host_ctxt         679 arch/arm64/kvm/hyp/switch.c 	__sysreg_save_state_nvhe(host_ctxt);
host_ctxt         699 arch/arm64/kvm/hyp/switch.c 		exit_code = __guest_enter(vcpu, host_ctxt);
host_ctxt         714 arch/arm64/kvm/hyp/switch.c 	__sysreg_restore_state_nvhe(host_ctxt);
host_ctxt         726 arch/arm64/kvm/hyp/switch.c 		__pmu_switch_to_host(host_ctxt);
host_ctxt         766 arch/arm64/kvm/hyp/switch.c 				 struct kvm_cpu_context *host_ctxt)
host_ctxt         769 arch/arm64/kvm/hyp/switch.c 	vcpu = host_ctxt->__hyp_running_vcpu;
host_ctxt         772 arch/arm64/kvm/hyp/switch.c 	sysreg_restore_host_state_vhe(host_ctxt);
host_ctxt         781 arch/arm64/kvm/hyp/switch.c void __hyp_text __noreturn hyp_panic(struct kvm_cpu_context *host_ctxt)
host_ctxt         788 arch/arm64/kvm/hyp/switch.c 		__hyp_call_panic_nvhe(spsr, elr, par, host_ctxt);
host_ctxt         790 arch/arm64/kvm/hyp/switch.c 		__hyp_call_panic_vhe(spsr, elr, par, host_ctxt);
host_ctxt         248 arch/arm64/kvm/hyp/sysreg-sr.c 	struct kvm_cpu_context *host_ctxt = vcpu->arch.host_cpu_context;
host_ctxt         254 arch/arm64/kvm/hyp/sysreg-sr.c 	__sysreg_save_user_state(host_ctxt);
host_ctxt         284 arch/arm64/kvm/hyp/sysreg-sr.c 	struct kvm_cpu_context *host_ctxt = vcpu->arch.host_cpu_context;
host_ctxt         297 arch/arm64/kvm/hyp/sysreg-sr.c 	__sysreg_restore_user_state(host_ctxt);
host_ctxt         166 arch/arm64/kvm/pmu.c 	struct kvm_cpu_context *host_ctxt;
host_ctxt         173 arch/arm64/kvm/pmu.c 	host_ctxt = vcpu->arch.host_cpu_context;
host_ctxt         174 arch/arm64/kvm/pmu.c 	host = container_of(host_ctxt, struct kvm_host_data, host_ctxt);
host_ctxt         187 arch/arm64/kvm/pmu.c 	struct kvm_cpu_context *host_ctxt;
host_ctxt         194 arch/arm64/kvm/pmu.c 	host_ctxt = vcpu->arch.host_cpu_context;
host_ctxt         195 arch/arm64/kvm/pmu.c 	host = container_of(host_ctxt, struct kvm_host_data, host_ctxt);
host_ctxt         385 virt/kvm/arm/arm.c 	vcpu->arch.host_cpu_context = &cpu_data->host_ctxt;
host_ctxt        1365 virt/kvm/arm/arm.c 	kvm_init_host_cpu_context(&this_cpu_ptr(&kvm_host_data)->host_ctxt);