Lines Matching refs:vmcb

95 	struct vmcb *hsave;
98 u64 vmcb; member
131 struct vmcb *vmcb; member
233 static inline void mark_all_dirty(struct vmcb *vmcb) in mark_all_dirty() argument
235 vmcb->control.clean = 0; in mark_all_dirty()
238 static inline void mark_all_clean(struct vmcb *vmcb) in mark_all_clean() argument
240 vmcb->control.clean = ((1 << VMCB_DIRTY_MAX) - 1) in mark_all_clean()
244 static inline void mark_dirty(struct vmcb *vmcb, int bit) in mark_dirty() argument
246 vmcb->control.clean &= ~(1 << bit); in mark_dirty()
259 mark_dirty(svm->vmcb, VMCB_INTERCEPTS); in recalc_intercepts()
264 c = &svm->vmcb->control; in recalc_intercepts()
274 static inline struct vmcb *get_host_vmcb(struct vcpu_svm *svm) in get_host_vmcb()
279 return svm->vmcb; in get_host_vmcb()
284 struct vmcb *vmcb = get_host_vmcb(svm); in set_cr_intercept() local
286 vmcb->control.intercept_cr |= (1U << bit); in set_cr_intercept()
293 struct vmcb *vmcb = get_host_vmcb(svm); in clr_cr_intercept() local
295 vmcb->control.intercept_cr &= ~(1U << bit); in clr_cr_intercept()
302 struct vmcb *vmcb = get_host_vmcb(svm); in is_cr_intercept() local
304 return vmcb->control.intercept_cr & (1U << bit); in is_cr_intercept()
309 struct vmcb *vmcb = get_host_vmcb(svm); in set_dr_intercepts() local
311 vmcb->control.intercept_dr = (1 << INTERCEPT_DR0_READ) in set_dr_intercepts()
333 struct vmcb *vmcb = get_host_vmcb(svm); in clr_dr_intercepts() local
335 vmcb->control.intercept_dr = 0; in clr_dr_intercepts()
342 struct vmcb *vmcb = get_host_vmcb(svm); in set_exception_intercept() local
344 vmcb->control.intercept_exceptions |= (1U << bit); in set_exception_intercept()
351 struct vmcb *vmcb = get_host_vmcb(svm); in clr_exception_intercept() local
353 vmcb->control.intercept_exceptions &= ~(1U << bit); in clr_exception_intercept()
360 struct vmcb *vmcb = get_host_vmcb(svm); in set_intercept() local
362 vmcb->control.intercept |= (1ULL << bit); in set_intercept()
369 struct vmcb *vmcb = get_host_vmcb(svm); in clr_intercept() local
371 vmcb->control.intercept &= ~(1ULL << bit); in clr_intercept()
479 to_svm(vcpu)->vmcb->save.efer = efer | EFER_SVME; in svm_set_efer()
480 mark_dirty(to_svm(vcpu)->vmcb, VMCB_CR); in svm_set_efer()
494 if (svm->vmcb->control.int_state & SVM_INTERRUPT_SHADOW_MASK) in svm_get_interrupt_shadow()
504 svm->vmcb->control.int_state &= ~SVM_INTERRUPT_SHADOW_MASK; in svm_set_interrupt_shadow()
506 svm->vmcb->control.int_state |= SVM_INTERRUPT_SHADOW_MASK; in svm_set_interrupt_shadow()
514 if (svm->vmcb->control.next_rip != 0) { in skip_emulated_instruction()
516 svm->next_rip = svm->vmcb->control.next_rip; in skip_emulated_instruction()
559 svm->int3_rip = rip + svm->vmcb->save.cs.base; in svm_queue_exception()
563 svm->vmcb->control.event_inj = nr in svm_queue_exception()
567 svm->vmcb->control.event_inj_err = error_code; in svm_queue_exception()
849 svm->vmcb->control.lbr_ctl = 1; in svm_enable_lbrv()
860 svm->vmcb->control.lbr_ctl = 0; in svm_disable_lbrv()
1035 return svm->vmcb->control.tsc_offset; in svm_read_tsc_offset()
1044 g_tsc_offset = svm->vmcb->control.tsc_offset - in svm_write_tsc_offset()
1049 svm->vmcb->control.tsc_offset, in svm_write_tsc_offset()
1052 svm->vmcb->control.tsc_offset = offset + g_tsc_offset; in svm_write_tsc_offset()
1054 mark_dirty(svm->vmcb, VMCB_INTERCEPTS); in svm_write_tsc_offset()
1067 svm->vmcb->control.tsc_offset += adjustment; in svm_adjust_tsc_offset()
1072 svm->vmcb->control.tsc_offset - adjustment, in svm_adjust_tsc_offset()
1073 svm->vmcb->control.tsc_offset); in svm_adjust_tsc_offset()
1075 mark_dirty(svm->vmcb, VMCB_INTERCEPTS); in svm_adjust_tsc_offset()
1089 struct vmcb_control_area *control = &svm->vmcb->control; in init_vmcb()
1090 struct vmcb_save_area *save = &svm->vmcb->save; in init_vmcb()
1189 svm->nested.vmcb = 0; in init_vmcb()
1197 mark_all_dirty(svm->vmcb); in init_vmcb()
1260 svm->vmcb = page_address(page); in svm_create_vcpu()
1261 clear_page(svm->vmcb); in svm_create_vcpu()
1308 mark_all_dirty(svm->vmcb); in svm_vcpu_load()
1350 return to_svm(vcpu)->vmcb->save.rflags; in svm_get_rflags()
1360 to_svm(vcpu)->vmcb->save.rflags = rflags; in svm_set_rflags()
1387 struct vmcb_save_area *save = &to_svm(vcpu)->vmcb->save; in svm_seg()
1473 var->dpl = to_svm(vcpu)->vmcb->save.cpl; in svm_get_segment()
1480 struct vmcb_save_area *save = &to_svm(vcpu)->vmcb->save; in svm_get_cpl()
1489 dt->size = svm->vmcb->save.idtr.limit; in svm_get_idt()
1490 dt->address = svm->vmcb->save.idtr.base; in svm_get_idt()
1497 svm->vmcb->save.idtr.limit = dt->size; in svm_set_idt()
1498 svm->vmcb->save.idtr.base = dt->address ; in svm_set_idt()
1499 mark_dirty(svm->vmcb, VMCB_DT); in svm_set_idt()
1506 dt->size = svm->vmcb->save.gdtr.limit; in svm_get_gdt()
1507 dt->address = svm->vmcb->save.gdtr.base; in svm_get_gdt()
1514 svm->vmcb->save.gdtr.limit = dt->size; in svm_set_gdt()
1515 svm->vmcb->save.gdtr.base = dt->address ; in svm_set_gdt()
1516 mark_dirty(svm->vmcb, VMCB_DT); in svm_set_gdt()
1534 u64 *hcr0 = &svm->vmcb->save.cr0; in update_cr0_intercept()
1542 mark_dirty(svm->vmcb, VMCB_CR); in update_cr0_intercept()
1561 svm->vmcb->save.efer |= EFER_LMA | EFER_LME; in svm_set_cr0()
1566 svm->vmcb->save.efer &= ~(EFER_LMA | EFER_LME); in svm_set_cr0()
1583 svm->vmcb->save.cr0 = cr0; in svm_set_cr0()
1584 mark_dirty(svm->vmcb, VMCB_CR); in svm_set_cr0()
1591 unsigned long old_cr4 = to_svm(vcpu)->vmcb->save.cr4; in svm_set_cr4()
1603 to_svm(vcpu)->vmcb->save.cr4 = cr4; in svm_set_cr4()
1604 mark_dirty(to_svm(vcpu)->vmcb, VMCB_CR); in svm_set_cr4()
1637 svm->vmcb->save.cpl = (s->attrib >> SVM_SELECTOR_DPL_SHIFT) & 3; in svm_set_segment()
1639 mark_dirty(svm->vmcb, VMCB_SEG); in svm_set_segment()
1660 svm->vmcb->control.tlb_ctl = TLB_CONTROL_FLUSH_ALL_ASID; in new_asid()
1664 svm->vmcb->control.asid = sd->next_asid++; in new_asid()
1666 mark_dirty(svm->vmcb, VMCB_ASID); in new_asid()
1671 return to_svm(vcpu)->vmcb->save.dr6; in svm_get_dr6()
1678 svm->vmcb->save.dr6 = value; in svm_set_dr6()
1679 mark_dirty(svm->vmcb, VMCB_DR); in svm_set_dr6()
1691 vcpu->arch.dr7 = svm->vmcb->save.dr7; in svm_sync_dirty_debug_regs()
1701 svm->vmcb->save.dr7 = value; in svm_set_dr7()
1702 mark_dirty(svm->vmcb, VMCB_DR); in svm_set_dr7()
1707 u64 fault_address = svm->vmcb->control.exit_info_2; in pf_interception()
1713 error_code = svm->vmcb->control.exit_info_1; in pf_interception()
1719 svm->vmcb->control.insn_bytes, in pf_interception()
1720 svm->vmcb->control.insn_len); in pf_interception()
1752 svm->vmcb->save.rflags &= in db_interception()
1760 svm->vmcb->save.cs.base + svm->vmcb->save.rip; in db_interception()
1773 kvm_run->debug.arch.pc = svm->vmcb->save.cs.base + svm->vmcb->save.rip; in bp_interception()
1887 clear_page(svm->vmcb); in shutdown_interception()
1897 u32 io_info = svm->vmcb->control.exit_info_1; /* address size bug? */ in io_interception()
1909 svm->next_rip = svm->vmcb->control.exit_info_2; in io_interception()
1970 svm->vmcb->control.nested_cr3 = root; in nested_svm_set_tdp_cr3()
1971 mark_dirty(svm->vmcb, VMCB_NPT); in nested_svm_set_tdp_cr3()
1980 if (svm->vmcb->control.exit_code != SVM_EXIT_NPF) { in nested_svm_inject_npf_exit()
1985 svm->vmcb->control.exit_code = SVM_EXIT_NPF; in nested_svm_inject_npf_exit()
1986 svm->vmcb->control.exit_code_hi = 0; in nested_svm_inject_npf_exit()
1987 svm->vmcb->control.exit_info_1 = (1ULL << 32); in nested_svm_inject_npf_exit()
1988 svm->vmcb->control.exit_info_2 = fault->address; in nested_svm_inject_npf_exit()
1991 svm->vmcb->control.exit_info_1 &= ~0xffffffffULL; in nested_svm_inject_npf_exit()
1992 svm->vmcb->control.exit_info_1 |= fault->error_code; in nested_svm_inject_npf_exit()
1998 if (svm->vmcb->control.exit_info_1 & (2ULL << 32)) in nested_svm_inject_npf_exit()
1999 svm->vmcb->control.exit_info_1 &= ~1; in nested_svm_inject_npf_exit()
2029 if (svm->vmcb->save.cpl) { in nested_svm_check_permissions()
2045 svm->vmcb->control.exit_code = SVM_EXIT_EXCP_BASE + nr; in nested_svm_check_exception()
2046 svm->vmcb->control.exit_code_hi = 0; in nested_svm_check_exception()
2047 svm->vmcb->control.exit_info_1 = error_code; in nested_svm_check_exception()
2048 svm->vmcb->control.exit_info_2 = svm->vcpu.arch.cr2; in nested_svm_check_exception()
2077 svm->vmcb->control.exit_code = SVM_EXIT_INTR; in nested_svm_intr()
2078 svm->vmcb->control.exit_info_1 = 0; in nested_svm_intr()
2079 svm->vmcb->control.exit_info_2 = 0; in nested_svm_intr()
2089 trace_kvm_nested_intr_vmexit(svm->vmcb->save.rip); in nested_svm_intr()
2105 svm->vmcb->control.exit_code = SVM_EXIT_NMI; in nested_svm_nmi()
2147 port = svm->vmcb->control.exit_info_1 >> 16; in nested_svm_intercept_ioio()
2148 size = (svm->vmcb->control.exit_info_1 & SVM_IOIO_SIZE_MASK) >> in nested_svm_intercept_ioio()
2172 write = svm->vmcb->control.exit_info_1 & 1; in nested_svm_exit_handled_msr()
2189 u32 exit_code = svm->vmcb->control.exit_code; in nested_svm_exit_special()
2221 u32 exit_code = svm->vmcb->control.exit_code; in nested_svm_intercept()
2279 static inline void copy_vmcb_control_area(struct vmcb *dst_vmcb, struct vmcb *from_vmcb) in copy_vmcb_control_area()
2311 struct vmcb *nested_vmcb; in nested_svm_vmexit()
2312 struct vmcb *hsave = svm->nested.hsave; in nested_svm_vmexit()
2313 struct vmcb *vmcb = svm->vmcb; in nested_svm_vmexit() local
2316 trace_kvm_nested_vmexit_inject(vmcb->control.exit_code, in nested_svm_vmexit()
2317 vmcb->control.exit_info_1, in nested_svm_vmexit()
2318 vmcb->control.exit_info_2, in nested_svm_vmexit()
2319 vmcb->control.exit_int_info, in nested_svm_vmexit()
2320 vmcb->control.exit_int_info_err, in nested_svm_vmexit()
2323 nested_vmcb = nested_svm_map(svm, svm->nested.vmcb, &page); in nested_svm_vmexit()
2329 svm->nested.vmcb = 0; in nested_svm_vmexit()
2334 nested_vmcb->save.es = vmcb->save.es; in nested_svm_vmexit()
2335 nested_vmcb->save.cs = vmcb->save.cs; in nested_svm_vmexit()
2336 nested_vmcb->save.ss = vmcb->save.ss; in nested_svm_vmexit()
2337 nested_vmcb->save.ds = vmcb->save.ds; in nested_svm_vmexit()
2338 nested_vmcb->save.gdtr = vmcb->save.gdtr; in nested_svm_vmexit()
2339 nested_vmcb->save.idtr = vmcb->save.idtr; in nested_svm_vmexit()
2343 nested_vmcb->save.cr2 = vmcb->save.cr2; in nested_svm_vmexit()
2346 nested_vmcb->save.rip = vmcb->save.rip; in nested_svm_vmexit()
2347 nested_vmcb->save.rsp = vmcb->save.rsp; in nested_svm_vmexit()
2348 nested_vmcb->save.rax = vmcb->save.rax; in nested_svm_vmexit()
2349 nested_vmcb->save.dr7 = vmcb->save.dr7; in nested_svm_vmexit()
2350 nested_vmcb->save.dr6 = vmcb->save.dr6; in nested_svm_vmexit()
2351 nested_vmcb->save.cpl = vmcb->save.cpl; in nested_svm_vmexit()
2353 nested_vmcb->control.int_ctl = vmcb->control.int_ctl; in nested_svm_vmexit()
2354 nested_vmcb->control.int_vector = vmcb->control.int_vector; in nested_svm_vmexit()
2355 nested_vmcb->control.int_state = vmcb->control.int_state; in nested_svm_vmexit()
2356 nested_vmcb->control.exit_code = vmcb->control.exit_code; in nested_svm_vmexit()
2357 nested_vmcb->control.exit_code_hi = vmcb->control.exit_code_hi; in nested_svm_vmexit()
2358 nested_vmcb->control.exit_info_1 = vmcb->control.exit_info_1; in nested_svm_vmexit()
2359 nested_vmcb->control.exit_info_2 = vmcb->control.exit_info_2; in nested_svm_vmexit()
2360 nested_vmcb->control.exit_int_info = vmcb->control.exit_int_info; in nested_svm_vmexit()
2361 nested_vmcb->control.exit_int_info_err = vmcb->control.exit_int_info_err; in nested_svm_vmexit()
2362 nested_vmcb->control.next_rip = vmcb->control.next_rip; in nested_svm_vmexit()
2372 if (vmcb->control.event_inj & SVM_EVTINJ_VALID) { in nested_svm_vmexit()
2375 nc->exit_int_info = vmcb->control.event_inj; in nested_svm_vmexit()
2376 nc->exit_int_info_err = vmcb->control.event_inj_err; in nested_svm_vmexit()
2388 copy_vmcb_control_area(vmcb, hsave); in nested_svm_vmexit()
2396 svm->vmcb->save.es = hsave->save.es; in nested_svm_vmexit()
2397 svm->vmcb->save.cs = hsave->save.cs; in nested_svm_vmexit()
2398 svm->vmcb->save.ss = hsave->save.ss; in nested_svm_vmexit()
2399 svm->vmcb->save.ds = hsave->save.ds; in nested_svm_vmexit()
2400 svm->vmcb->save.gdtr = hsave->save.gdtr; in nested_svm_vmexit()
2401 svm->vmcb->save.idtr = hsave->save.idtr; in nested_svm_vmexit()
2407 svm->vmcb->save.cr3 = hsave->save.cr3; in nested_svm_vmexit()
2415 svm->vmcb->save.dr7 = 0; in nested_svm_vmexit()
2416 svm->vmcb->save.cpl = 0; in nested_svm_vmexit()
2417 svm->vmcb->control.exit_int_info = 0; in nested_svm_vmexit()
2419 mark_all_dirty(svm->vmcb); in nested_svm_vmexit()
2458 svm->vmcb->control.msrpm_base_pa = __pa(svm->nested.msrpm); in nested_svm_vmrun_msrpm()
2463 static bool nested_vmcb_checks(struct vmcb *vmcb) in nested_vmcb_checks() argument
2465 if ((vmcb->control.intercept & (1ULL << INTERCEPT_VMRUN)) == 0) in nested_vmcb_checks()
2468 if (vmcb->control.asid == 0) in nested_vmcb_checks()
2471 if (vmcb->control.nested_ctl && !npt_enabled) in nested_vmcb_checks()
2479 struct vmcb *nested_vmcb; in nested_svm_vmrun()
2480 struct vmcb *hsave = svm->nested.hsave; in nested_svm_vmrun()
2481 struct vmcb *vmcb = svm->vmcb; in nested_svm_vmrun() local
2485 vmcb_gpa = svm->vmcb->save.rax; in nested_svm_vmrun()
2487 nested_vmcb = nested_svm_map(svm, svm->vmcb->save.rax, &page); in nested_svm_vmrun()
2502 trace_kvm_nested_vmrun(svm->vmcb->save.rip, vmcb_gpa, in nested_svm_vmrun()
2521 hsave->save.es = vmcb->save.es; in nested_svm_vmrun()
2522 hsave->save.cs = vmcb->save.cs; in nested_svm_vmrun()
2523 hsave->save.ss = vmcb->save.ss; in nested_svm_vmrun()
2524 hsave->save.ds = vmcb->save.ds; in nested_svm_vmrun()
2525 hsave->save.gdtr = vmcb->save.gdtr; in nested_svm_vmrun()
2526 hsave->save.idtr = vmcb->save.idtr; in nested_svm_vmrun()
2532 hsave->save.rsp = vmcb->save.rsp; in nested_svm_vmrun()
2533 hsave->save.rax = vmcb->save.rax; in nested_svm_vmrun()
2535 hsave->save.cr3 = vmcb->save.cr3; in nested_svm_vmrun()
2539 copy_vmcb_control_area(hsave, vmcb); in nested_svm_vmrun()
2553 svm->vmcb->save.es = nested_vmcb->save.es; in nested_svm_vmrun()
2554 svm->vmcb->save.cs = nested_vmcb->save.cs; in nested_svm_vmrun()
2555 svm->vmcb->save.ss = nested_vmcb->save.ss; in nested_svm_vmrun()
2556 svm->vmcb->save.ds = nested_vmcb->save.ds; in nested_svm_vmrun()
2557 svm->vmcb->save.gdtr = nested_vmcb->save.gdtr; in nested_svm_vmrun()
2558 svm->vmcb->save.idtr = nested_vmcb->save.idtr; in nested_svm_vmrun()
2564 svm->vmcb->save.cr3 = nested_vmcb->save.cr3; in nested_svm_vmrun()
2572 svm->vmcb->save.cr2 = svm->vcpu.arch.cr2 = nested_vmcb->save.cr2; in nested_svm_vmrun()
2578 svm->vmcb->save.rax = nested_vmcb->save.rax; in nested_svm_vmrun()
2579 svm->vmcb->save.rsp = nested_vmcb->save.rsp; in nested_svm_vmrun()
2580 svm->vmcb->save.rip = nested_vmcb->save.rip; in nested_svm_vmrun()
2581 svm->vmcb->save.dr7 = nested_vmcb->save.dr7; in nested_svm_vmrun()
2582 svm->vmcb->save.dr6 = nested_vmcb->save.dr6; in nested_svm_vmrun()
2583 svm->vmcb->save.cpl = nested_vmcb->save.cpl; in nested_svm_vmrun()
2595 svm->vmcb->control.int_ctl = nested_vmcb->control.int_ctl | V_INTR_MASKING_MASK; in nested_svm_vmrun()
2610 svm->vmcb->control.lbr_ctl = nested_vmcb->control.lbr_ctl; in nested_svm_vmrun()
2611 svm->vmcb->control.int_vector = nested_vmcb->control.int_vector; in nested_svm_vmrun()
2612 svm->vmcb->control.int_state = nested_vmcb->control.int_state; in nested_svm_vmrun()
2613 svm->vmcb->control.tsc_offset += nested_vmcb->control.tsc_offset; in nested_svm_vmrun()
2614 svm->vmcb->control.event_inj = nested_vmcb->control.event_inj; in nested_svm_vmrun()
2615 svm->vmcb->control.event_inj_err = nested_vmcb->control.event_inj_err; in nested_svm_vmrun()
2628 svm->nested.vmcb = vmcb_gpa; in nested_svm_vmrun()
2632 mark_all_dirty(svm->vmcb); in nested_svm_vmrun()
2637 static void nested_svm_vmloadsave(struct vmcb *from_vmcb, struct vmcb *to_vmcb) in nested_svm_vmloadsave()
2655 struct vmcb *nested_vmcb; in vmload_interception()
2661 nested_vmcb = nested_svm_map(svm, svm->vmcb->save.rax, &page); in vmload_interception()
2668 nested_svm_vmloadsave(nested_vmcb, svm->vmcb); in vmload_interception()
2676 struct vmcb *nested_vmcb; in vmsave_interception()
2682 nested_vmcb = nested_svm_map(svm, svm->vmcb->save.rax, &page); in vmsave_interception()
2689 nested_svm_vmloadsave(svm->vmcb, nested_vmcb); in vmsave_interception()
2713 svm->vmcb->control.exit_code = SVM_EXIT_ERR; in vmrun_interception()
2714 svm->vmcb->control.exit_code_hi = 0; in vmrun_interception()
2715 svm->vmcb->control.exit_info_1 = 0; in vmrun_interception()
2716 svm->vmcb->control.exit_info_2 = 0; in vmrun_interception()
2749 svm->vmcb->control.int_ctl &= ~V_IRQ_MASK; in clgi_interception()
2751 mark_dirty(svm->vmcb, VMCB_INTR); in clgi_interception()
2760 trace_kvm_invlpga(svm->vmcb->save.rip, kvm_register_read(&svm->vcpu, VCPU_REGS_RCX), in invlpga_interception()
2773 trace_kvm_skinit(svm->vmcb->save.rip, kvm_register_read(&svm->vcpu, VCPU_REGS_RAX)); in skinit_interception()
2802 int int_type = svm->vmcb->control.exit_int_info & in task_switch_interception()
2804 int int_vec = svm->vmcb->control.exit_int_info & SVM_EVTINJ_VEC_MASK; in task_switch_interception()
2806 svm->vmcb->control.exit_int_info & SVM_EXITINTINFO_TYPE_MASK; in task_switch_interception()
2808 svm->vmcb->control.exit_int_info & SVM_EXITINTINFO_VALID; in task_switch_interception()
2812 tss_selector = (u16)svm->vmcb->control.exit_info_1; in task_switch_interception()
2814 if (svm->vmcb->control.exit_info_2 & in task_switch_interception()
2817 else if (svm->vmcb->control.exit_info_2 & in task_switch_interception()
2831 if (svm->vmcb->control.exit_info_2 & in task_switch_interception()
2835 (u32)svm->vmcb->control.exit_info_2; in task_switch_interception()
2888 kvm_mmu_invlpg(&svm->vcpu, svm->vmcb->control.exit_info_1); in invlpg_interception()
2928 svm->vmcb->control.exit_code = SVM_EXIT_CR0_SEL_WRITE; in check_selective_cr0_intercepted()
2946 if (unlikely((svm->vmcb->control.exit_info_1 & CR_VALID) == 0)) in cr_interception()
2949 reg = svm->vmcb->control.exit_info_1 & SVM_EXITINFO_REG_MASK; in cr_interception()
2950 if (svm->vmcb->control.exit_code == SVM_EXIT_CR0_SEL_WRITE) in cr_interception()
2953 cr = svm->vmcb->control.exit_code - SVM_EXIT_READ_CR0; in cr_interception()
3029 reg = svm->vmcb->control.exit_info_1 & SVM_EXITINFO_REG_MASK; in dr_interception()
3030 dr = svm->vmcb->control.exit_code - SVM_EXIT_READ_DR0; in dr_interception()
3067 struct vmcb *vmcb = get_host_vmcb(to_svm(vcpu)); in svm_read_l1_tsc() local
3068 return vmcb->control.tsc_offset + in svm_read_l1_tsc()
3078 *data = svm->vmcb->control.tsc_offset + in svm_get_msr()
3084 *data = svm->vmcb->save.star; in svm_get_msr()
3088 *data = svm->vmcb->save.lstar; in svm_get_msr()
3091 *data = svm->vmcb->save.cstar; in svm_get_msr()
3094 *data = svm->vmcb->save.kernel_gs_base; in svm_get_msr()
3097 *data = svm->vmcb->save.sfmask; in svm_get_msr()
3101 *data = svm->vmcb->save.sysenter_cs; in svm_get_msr()
3115 *data = svm->vmcb->save.dbgctl; in svm_get_msr()
3118 *data = svm->vmcb->save.br_from; in svm_get_msr()
3121 *data = svm->vmcb->save.br_to; in svm_get_msr()
3124 *data = svm->vmcb->save.last_excp_from; in svm_get_msr()
3127 *data = svm->vmcb->save.last_excp_to; in svm_get_msr()
3199 svm->vmcb->save.star = data; in svm_set_msr()
3203 svm->vmcb->save.lstar = data; in svm_set_msr()
3206 svm->vmcb->save.cstar = data; in svm_set_msr()
3209 svm->vmcb->save.kernel_gs_base = data; in svm_set_msr()
3212 svm->vmcb->save.sfmask = data; in svm_set_msr()
3216 svm->vmcb->save.sysenter_cs = data; in svm_set_msr()
3220 svm->vmcb->save.sysenter_eip = data; in svm_set_msr()
3224 svm->vmcb->save.sysenter_esp = data; in svm_set_msr()
3235 svm->vmcb->save.dbgctl = data; in svm_set_msr()
3236 mark_dirty(svm->vmcb, VMCB_LBR); in svm_set_msr()
3279 if (svm->vmcb->control.exit_info_1) in msr_interception()
3291 svm->vmcb->control.int_ctl &= ~V_IRQ_MASK; in interrupt_window_interception()
3292 mark_dirty(svm->vmcb, VMCB_INTR); in interrupt_window_interception()
3399 struct vmcb_control_area *control = &svm->vmcb->control; in dump_vmcb()
3400 struct vmcb_save_area *save = &svm->vmcb->save; in dump_vmcb()
3503 struct vmcb_control_area *control = &to_svm(vcpu)->vmcb->control; in svm_get_exit_info()
3513 u32 exit_code = svm->vmcb->control.exit_code; in handle_exit()
3516 vcpu->arch.cr0 = svm->vmcb->save.cr0; in handle_exit()
3518 vcpu->arch.cr3 = svm->vmcb->save.cr3; in handle_exit()
3530 trace_kvm_nested_vmexit(svm->vmcb->save.rip, exit_code, in handle_exit()
3531 svm->vmcb->control.exit_info_1, in handle_exit()
3532 svm->vmcb->control.exit_info_2, in handle_exit()
3533 svm->vmcb->control.exit_int_info, in handle_exit()
3534 svm->vmcb->control.exit_int_info_err, in handle_exit()
3548 if (svm->vmcb->control.exit_code == SVM_EXIT_ERR) { in handle_exit()
3551 = svm->vmcb->control.exit_code; in handle_exit()
3557 if (is_external_interrupt(svm->vmcb->control.exit_int_info) && in handle_exit()
3563 __func__, svm->vmcb->control.exit_int_info, in handle_exit()
3600 svm->vmcb->control.event_inj = SVM_EVTINJ_VALID | SVM_EVTINJ_TYPE_NMI; in svm_inject_nmi()
3610 control = &svm->vmcb->control; in svm_inject_irq()
3615 mark_dirty(svm->vmcb, VMCB_INTR); in svm_inject_irq()
3627 svm->vmcb->control.event_inj = vcpu->arch.interrupt.nr | in svm_set_irq()
3670 struct vmcb *vmcb = svm->vmcb; in svm_nmi_allowed() local
3672 ret = !(vmcb->control.int_state & SVM_INTERRUPT_SHADOW_MASK) && in svm_nmi_allowed()
3702 struct vmcb *vmcb = svm->vmcb; in svm_interrupt_allowed() local
3706 (vmcb->control.int_state & SVM_INTERRUPT_SHADOW_MASK)) in svm_interrupt_allowed()
3746 svm->vmcb->save.rflags |= (X86_EFLAGS_TF | X86_EFLAGS_RF); in enable_nmi_window()
3759 svm->vmcb->control.tlb_ctl = TLB_CONTROL_FLUSH_ASID; in svm_flush_tlb()
3776 int cr8 = svm->vmcb->control.int_ctl & V_TPR_MASK; in sync_cr8_to_lapic()
3790 svm->vmcb->control.int_ctl &= ~V_TPR_MASK; in sync_lapic_to_cr8()
3791 svm->vmcb->control.int_ctl |= cr8 & V_TPR_MASK; in sync_lapic_to_cr8()
3798 u32 exitintinfo = svm->vmcb->control.exit_int_info; in svm_complete_interrupts()
3844 u32 err = svm->vmcb->control.exit_int_info_err; in svm_complete_interrupts()
3861 struct vmcb_control_area *control = &svm->vmcb->control; in svm_cancel_injection()
3873 svm->vmcb->save.rax = vcpu->arch.regs[VCPU_REGS_RAX]; in svm_vcpu_run()
3874 svm->vmcb->save.rsp = vcpu->arch.regs[VCPU_REGS_RSP]; in svm_vcpu_run()
3875 svm->vmcb->save.rip = vcpu->arch.regs[VCPU_REGS_RIP]; in svm_vcpu_run()
3888 svm->vmcb->save.cr2 = vcpu->arch.cr2; in svm_vcpu_run()
3941 [vmcb]"i"(offsetof(struct vcpu_svm, vmcb_pa)), in svm_vcpu_run()
3980 vcpu->arch.cr2 = svm->vmcb->save.cr2; in svm_vcpu_run()
3981 vcpu->arch.regs[VCPU_REGS_RAX] = svm->vmcb->save.rax; in svm_vcpu_run()
3982 vcpu->arch.regs[VCPU_REGS_RSP] = svm->vmcb->save.rsp; in svm_vcpu_run()
3983 vcpu->arch.regs[VCPU_REGS_RIP] = svm->vmcb->save.rip; in svm_vcpu_run()
3985 trace_kvm_exit(svm->vmcb->control.exit_code, vcpu, KVM_ISA_SVM); in svm_vcpu_run()
3987 if (unlikely(svm->vmcb->control.exit_code == SVM_EXIT_NMI)) in svm_vcpu_run()
3994 if (unlikely(svm->vmcb->control.exit_code == SVM_EXIT_NMI)) in svm_vcpu_run()
4001 svm->vmcb->control.tlb_ctl = TLB_CONTROL_DO_NOTHING; in svm_vcpu_run()
4004 if (svm->vmcb->control.exit_code == SVM_EXIT_EXCP_BASE + PF_VECTOR) in svm_vcpu_run()
4016 if (unlikely(svm->vmcb->control.exit_code == in svm_vcpu_run()
4020 mark_all_clean(svm->vmcb); in svm_vcpu_run()
4027 svm->vmcb->save.cr3 = root; in svm_set_cr3()
4028 mark_dirty(svm->vmcb, VMCB_CR); in svm_set_cr3()
4036 svm->vmcb->control.nested_cr3 = root; in set_tdp_cr3()
4037 mark_dirty(svm->vmcb, VMCB_NPT); in set_tdp_cr3()
4040 svm->vmcb->save.cr3 = kvm_read_cr3(vcpu); in set_tdp_cr3()
4041 mark_dirty(svm->vmcb, VMCB_CR); in set_tdp_cr3()
4222 struct vmcb *vmcb = svm->vmcb; in svm_check_intercept() local
4275 vmcb->control.exit_info_1 = 1; in svm_check_intercept()
4277 vmcb->control.exit_info_1 = 0; in svm_check_intercept()
4313 vmcb->control.exit_info_1 = exit_info; in svm_check_intercept()
4314 vmcb->control.exit_info_2 = info->next_rip; in svm_check_intercept()
4324 vmcb->control.next_rip = info->next_rip; in svm_check_intercept()
4325 vmcb->control.exit_code = icpt_info.exit_code; in svm_check_intercept()