vmcs              293 arch/x86/kvm/vmx/nested.c static void vmx_switch_vmcs(struct kvm_vcpu *vcpu, struct loaded_vmcs *vmcs)
vmcs              299 arch/x86/kvm/vmx/nested.c 	if (vmx->loaded_vmcs == vmcs)
vmcs              304 arch/x86/kvm/vmx/nested.c 	vmx->loaded_vmcs = vmcs;
vmcs             1368 arch/x86/kvm/vmx/nested.c 	struct vmcs *shadow_vmcs = vmx->vmcs01.shadow_vmcs;
vmcs             1388 arch/x86/kvm/vmx/nested.c 	vmcs_load(vmx->loaded_vmcs->vmcs);
vmcs             1403 arch/x86/kvm/vmx/nested.c 	struct vmcs *shadow_vmcs = vmx->vmcs01.shadow_vmcs;
vmcs             1424 arch/x86/kvm/vmx/nested.c 	vmcs_load(vmx->loaded_vmcs->vmcs);
vmcs             4354 arch/x86/kvm/vmx/nested.c static struct vmcs *alloc_shadow_vmcs(struct kvm_vcpu *vcpu)
vmcs             4788 arch/x86/kvm/vmx/nested.c 			vmcs_load(vmx->loaded_vmcs->vmcs);
vmcs               19 arch/x86/kvm/vmx/ops.h void vmclear_error(struct vmcs *vmcs, u64 phys_addr);
vmcs               20 arch/x86/kvm/vmx/ops.h void vmptrld_error(struct vmcs *vmcs, u64 phys_addr);
vmcs              234 arch/x86/kvm/vmx/ops.h static inline void vmcs_clear(struct vmcs *vmcs)
vmcs              236 arch/x86/kvm/vmx/ops.h 	u64 phys_addr = __pa(vmcs);
vmcs              238 arch/x86/kvm/vmx/ops.h 	vmx_asm1(vmclear, "m"(phys_addr), vmcs, phys_addr);
vmcs              241 arch/x86/kvm/vmx/ops.h static inline void vmcs_load(struct vmcs *vmcs)
vmcs              243 arch/x86/kvm/vmx/ops.h 	u64 phys_addr = __pa(vmcs);
vmcs              248 arch/x86/kvm/vmx/ops.h 	vmx_asm1(vmptrld, "m"(phys_addr), vmcs, phys_addr);
vmcs               25 arch/x86/kvm/vmx/vmcs.h DECLARE_PER_CPU(struct vmcs *, current_vmcs);
vmcs               59 arch/x86/kvm/vmx/vmcs.h 	struct vmcs *vmcs;
vmcs               60 arch/x86/kvm/vmx/vmcs.h 	struct vmcs *shadow_vmcs;
vmcs              371 arch/x86/kvm/vmx/vmx.c noinline void vmclear_error(struct vmcs *vmcs, u64 phys_addr)
vmcs              373 arch/x86/kvm/vmx/vmx.c 	vmx_insn_failed("kvm: vmclear failed: %p/%llx\n", vmcs, phys_addr);
vmcs              376 arch/x86/kvm/vmx/vmx.c noinline void vmptrld_error(struct vmcs *vmcs, u64 phys_addr)
vmcs              378 arch/x86/kvm/vmx/vmx.c 	vmx_insn_failed("kvm: vmptrld failed: %p/%llx\n", vmcs, phys_addr);
vmcs              393 arch/x86/kvm/vmx/vmx.c static DEFINE_PER_CPU(struct vmcs *, vmxarea);
vmcs              394 arch/x86/kvm/vmx/vmx.c DEFINE_PER_CPU(struct vmcs *, current_vmcs);
vmcs              551 arch/x86/kvm/vmx/vmx.c 	evmcs = (struct hv_enlightened_vmcs *)to_vmx(vcpu)->loaded_vmcs->vmcs;
vmcs              643 arch/x86/kvm/vmx/vmx.c 	vmcs_clear(loaded_vmcs->vmcs);
vmcs              658 arch/x86/kvm/vmx/vmx.c 		vmcs_clear(v->vmcs);
vmcs              669 arch/x86/kvm/vmx/vmx.c 	if (per_cpu(current_vmcs, cpu) == loaded_vmcs->vmcs)
vmcs              672 arch/x86/kvm/vmx/vmx.c 	vmcs_clear(loaded_vmcs->vmcs);
vmcs             1294 arch/x86/kvm/vmx/vmx.c 	struct vmcs *prev;
vmcs             1314 arch/x86/kvm/vmx/vmx.c 	if (prev != vmx->loaded_vmcs->vmcs) {
vmcs             1315 arch/x86/kvm/vmx/vmx.c 		per_cpu(current_vmcs, cpu) = vmx->loaded_vmcs->vmcs;
vmcs             1316 arch/x86/kvm/vmx/vmx.c 		vmcs_load(vmx->loaded_vmcs->vmcs);
vmcs             1323 arch/x86/kvm/vmx/vmx.c 		if (!buddy || WARN_ON_ONCE(buddy->vmcs != prev))
vmcs             2502 arch/x86/kvm/vmx/vmx.c struct vmcs *alloc_vmcs_cpu(bool shadow, int cpu, gfp_t flags)
vmcs             2506 arch/x86/kvm/vmx/vmx.c 	struct vmcs *vmcs;
vmcs             2511 arch/x86/kvm/vmx/vmx.c 	vmcs = page_address(pages);
vmcs             2512 arch/x86/kvm/vmx/vmx.c 	memset(vmcs, 0, vmcs_config.size);
vmcs             2516 arch/x86/kvm/vmx/vmx.c 		vmcs->hdr.revision_id = KVM_EVMCS_VERSION;
vmcs             2518 arch/x86/kvm/vmx/vmx.c 		vmcs->hdr.revision_id = vmcs_config.revision_id;
vmcs             2521 arch/x86/kvm/vmx/vmx.c 		vmcs->hdr.shadow_vmcs = 1;
vmcs             2522 arch/x86/kvm/vmx/vmx.c 	return vmcs;
vmcs             2525 arch/x86/kvm/vmx/vmx.c void free_vmcs(struct vmcs *vmcs)
vmcs             2527 arch/x86/kvm/vmx/vmx.c 	free_pages((unsigned long)vmcs, vmcs_config.order);
vmcs             2535 arch/x86/kvm/vmx/vmx.c 	if (!loaded_vmcs->vmcs)
vmcs             2538 arch/x86/kvm/vmx/vmx.c 	free_vmcs(loaded_vmcs->vmcs);
vmcs             2539 arch/x86/kvm/vmx/vmx.c 	loaded_vmcs->vmcs = NULL;
vmcs             2547 arch/x86/kvm/vmx/vmx.c 	loaded_vmcs->vmcs = alloc_vmcs(false);
vmcs             2548 arch/x86/kvm/vmx/vmx.c 	if (!loaded_vmcs->vmcs)
vmcs             2566 arch/x86/kvm/vmx/vmx.c 				(struct hv_enlightened_vmcs *)loaded_vmcs->vmcs;
vmcs             2598 arch/x86/kvm/vmx/vmx.c 		struct vmcs *vmcs;
vmcs             2600 arch/x86/kvm/vmx/vmx.c 		vmcs = alloc_vmcs_cpu(false, cpu, GFP_KERNEL);
vmcs             2601 arch/x86/kvm/vmx/vmx.c 		if (!vmcs) {
vmcs             2617 arch/x86/kvm/vmx/vmx.c 			vmcs->hdr.revision_id = vmcs_config.revision_id;
vmcs             2619 arch/x86/kvm/vmx/vmx.c 		per_cpu(vmxarea, cpu) = vmcs;
vmcs              476 arch/x86/kvm/vmx/vmx.h struct vmcs *alloc_vmcs_cpu(bool shadow, int cpu, gfp_t flags);
vmcs              477 arch/x86/kvm/vmx/vmx.h void free_vmcs(struct vmcs *vmcs);
vmcs              483 arch/x86/kvm/vmx/vmx.h static inline struct vmcs *alloc_vmcs(bool shadow)
vmcs              239 tools/testing/selftests/kvm/include/evmcs.h static inline int evmcs_vmptrld(uint64_t vmcs_pa, void *vmcs)
vmcs              244 tools/testing/selftests/kvm/include/evmcs.h 	current_evmcs = vmcs;
vmcs              547 tools/testing/selftests/kvm/include/x86_64/vmx.h 	void *vmcs;
vmcs               89 tools/testing/selftests/kvm/lib/x86_64/vmx.c 	vmx->vmcs = (void *)vm_vaddr_alloc(vm, getpagesize(), 0x10000, 0, 0);
vmcs               90 tools/testing/selftests/kvm/lib/x86_64/vmx.c 	vmx->vmcs_hva = addr_gva2hva(vm, (uintptr_t)vmx->vmcs);
vmcs               91 tools/testing/selftests/kvm/lib/x86_64/vmx.c 	vmx->vmcs_gpa = addr_gva2gpa(vm, (uintptr_t)vmx->vmcs);
vmcs              181 tools/testing/selftests/kvm/lib/x86_64/vmx.c 		*(uint32_t *)(vmx->vmcs) = vmcs_revision();