Lines Matching refs:arch
90 struct kvm_lapic *apic = vcpu->arch.apic; in kvm_apic_pending_eoi()
166 mutex_lock(&kvm->arch.apic_map_lock); in recalculate_apic_map()
172 struct kvm_lapic *apic = vcpu->arch.apic; in recalculate_apic_map()
204 old = rcu_dereference_protected(kvm->arch.apic_map, in recalculate_apic_map()
205 lockdep_is_held(&kvm->arch.apic_map_lock)); in recalculate_apic_map()
206 rcu_assign_pointer(kvm->arch.apic_map, new); in recalculate_apic_map()
207 mutex_unlock(&kvm->arch.apic_map_lock); in recalculate_apic_map()
284 struct kvm_lapic *apic = vcpu->arch.apic; in kvm_apic_set_version()
348 struct kvm_lapic *apic = vcpu->arch.apic; in kvm_apic_update_irr()
490 highest_irr = apic_find_highest_irr(vcpu->arch.apic); in kvm_lapic_find_highest_irr()
502 struct kvm_lapic *apic = vcpu->arch.apic; in kvm_apic_set_irq()
511 return kvm_write_guest_cached(vcpu->kvm, &vcpu->arch.pv_eoi.data, &val, in pv_eoi_put_user()
518 return kvm_read_guest_cached(vcpu->kvm, &vcpu->arch.pv_eoi.data, val, in pv_eoi_get_user()
524 return vcpu->arch.pv_eoi.msr_val & KVM_MSR_ENABLED; in pv_eoi_enabled()
532 (unsigned long long)vcpu->arch.pv_eoi.msr_val); in pv_eoi_get_pending()
540 (unsigned long long)vcpu->arch.pv_eoi.msr_val); in pv_eoi_set_pending()
543 __set_bit(KVM_APIC_PV_EOI_PENDING, &vcpu->arch.apic_attention); in pv_eoi_set_pending()
550 (unsigned long long)vcpu->arch.pv_eoi.msr_val); in pv_eoi_clr_pending()
553 __clear_bit(KVM_APIC_PV_EOI_PENDING, &vcpu->arch.apic_attention); in pv_eoi_clr_pending()
654 struct kvm_lapic *target = vcpu->arch.apic; in kvm_apic_match_dest()
706 map = rcu_dereference(kvm->arch.apic_map); in kvm_irq_delivery_to_apic_fast()
771 map = rcu_dereference(kvm->arch.apic_map); in kvm_intr_is_single_vcpu_fast()
834 vcpu->arch.apic_arb_prio++; in __apic_accept_irq()
867 vcpu->arch.pv.pv_unhalted = 1; in __apic_accept_irq()
930 return vcpu1->arch.apic_arb_prio - vcpu2->arch.apic_arb_prio; in kvm_apic_compare_prio()
935 return test_bit(vector, (ulong *)apic->vcpu->arch.eoi_exit_bitmap); in kvm_ioapic_handles_vector()
948 apic->vcpu->arch.pending_ioapic_eoi = vector; in kvm_ioapic_send_eoi()
988 struct kvm_lapic *apic = vcpu->arch.apic; in kvm_apic_set_eoi_accelerated()
1064 if (apic->vcpu->arch.tpr_access_reporting) in report_tpr_access()
1221 struct kvm_lapic *apic = vcpu->arch.apic; in lapic_timer_int_injected()
1239 struct kvm_lapic *apic = vcpu->arch.apic; in wait_lapic_expire()
1312 unsigned long this_tsc_khz = vcpu->arch.virtual_tsc_khz; in start_apic_timer()
1345 atomic_inc(&apic->vcpu->kvm->arch.vapics_in_nmi_mode); in apic_manage_nmi_watchdog()
1347 atomic_dec(&apic->vcpu->kvm->arch.vapics_in_nmi_mode); in apic_manage_nmi_watchdog()
1519 apic_reg_write(vcpu->arch.apic, APIC_EOI, 0); in kvm_lapic_set_eoi()
1531 apic_reg_read(vcpu->arch.apic, offset, 4, &val); in kvm_apic_write_nodecode()
1534 apic_reg_write(vcpu->arch.apic, offset, val); in kvm_apic_write_nodecode()
1540 struct kvm_lapic *apic = vcpu->arch.apic; in kvm_free_lapic()
1542 if (!vcpu->arch.apic) in kvm_free_lapic()
1547 if (!(vcpu->arch.apic_base & MSR_IA32_APICBASE_ENABLE)) in kvm_free_lapic()
1567 struct kvm_lapic *apic = vcpu->arch.apic; in kvm_get_lapic_tscdeadline_msr()
1578 struct kvm_lapic *apic = vcpu->arch.apic; in kvm_set_lapic_tscdeadline_msr()
1591 struct kvm_lapic *apic = vcpu->arch.apic; in kvm_lapic_set_tpr()
1607 tpr = (u64) kvm_apic_get_reg(vcpu->arch.apic, APIC_TASKPRI); in kvm_lapic_get_cr8()
1614 u64 old_value = vcpu->arch.apic_base; in kvm_lapic_set_base()
1615 struct kvm_lapic *apic = vcpu->arch.apic; in kvm_lapic_set_base()
1619 vcpu->arch.apic_base = value; in kvm_lapic_set_base()
1623 vcpu->arch.apic_base = value; in kvm_lapic_set_base()
1642 apic->base_address = apic->vcpu->arch.apic_base & in kvm_lapic_set_base()
1651 "0x%lx.\n", apic->vcpu->arch.apic_base, apic->base_address); in kvm_lapic_set_base()
1663 apic = vcpu->arch.apic; in kvm_lapic_reset()
1703 vcpu->arch.apic_base | MSR_IA32_APICBASE_BSP); in kvm_lapic_reset()
1704 vcpu->arch.pv_eoi.msr_val = 0; in kvm_lapic_reset()
1707 vcpu->arch.apic_arb_prio = 0; in kvm_lapic_reset()
1708 vcpu->arch.apic_attention = 0; in kvm_lapic_reset()
1713 vcpu->arch.apic_base, apic->base_address); in kvm_lapic_reset()
1729 struct kvm_lapic *apic = vcpu->arch.apic; in apic_has_pending_timer()
1755 struct kvm_lapic *apic = vcpu->arch.apic; in kvm_apic_nmi_wd_deliver()
1791 vcpu->arch.apic = apic; in kvm_create_lapic()
1809 vcpu->arch.apic_base = MSR_IA32_APICBASE_ENABLE; in kvm_create_lapic()
1826 struct kvm_lapic *apic = vcpu->arch.apic; in kvm_apic_has_interrupt()
1842 u32 lvt0 = kvm_apic_get_reg(vcpu->arch.apic, APIC_LVT0); in kvm_apic_accept_pic_intr()
1845 if (!kvm_apic_hw_enabled(vcpu->arch.apic)) in kvm_apic_accept_pic_intr()
1855 struct kvm_lapic *apic = vcpu->arch.apic; in kvm_inject_apic_timer_irqs()
1871 struct kvm_lapic *apic = vcpu->arch.apic; in kvm_get_apic_interrupt()
1892 struct kvm_lapic *apic = vcpu->arch.apic; in kvm_apic_post_state_restore()
1894 kvm_lapic_set_base(vcpu, vcpu->arch.apic_base); in kvm_apic_post_state_restore()
1897 memcpy(vcpu->arch.apic->regs, s->regs, sizeof *s); in kvm_apic_post_state_restore()
1922 vcpu->arch.apic_arb_prio = 0; in kvm_apic_post_state_restore()
1932 timer = &vcpu->arch.apic->lapic_timer.timer; in __kvm_migrate_apic_timer()
1978 if (test_bit(KVM_APIC_PV_EOI_PENDING, &vcpu->arch.apic_attention)) in kvm_lapic_sync_from_vapic()
1979 apic_sync_pv_eoi_from_guest(vcpu, vcpu->arch.apic); in kvm_lapic_sync_from_vapic()
1981 if (!test_bit(KVM_APIC_CHECK_VAPIC, &vcpu->arch.apic_attention)) in kvm_lapic_sync_from_vapic()
1984 if (kvm_read_guest_cached(vcpu->kvm, &vcpu->arch.apic->vapic_cache, &data, in kvm_lapic_sync_from_vapic()
1988 apic_set_tpr(vcpu->arch.apic, data & 0xff); in kvm_lapic_sync_from_vapic()
2021 struct kvm_lapic *apic = vcpu->arch.apic; in kvm_lapic_sync_to_vapic()
2025 if (!test_bit(KVM_APIC_CHECK_VAPIC, &vcpu->arch.apic_attention)) in kvm_lapic_sync_to_vapic()
2037 kvm_write_guest_cached(vcpu->kvm, &vcpu->arch.apic->vapic_cache, &data, in kvm_lapic_sync_to_vapic()
2045 &vcpu->arch.apic->vapic_cache, in kvm_lapic_set_vapic_addr()
2048 __set_bit(KVM_APIC_CHECK_VAPIC, &vcpu->arch.apic_attention); in kvm_lapic_set_vapic_addr()
2050 __clear_bit(KVM_APIC_CHECK_VAPIC, &vcpu->arch.apic_attention); in kvm_lapic_set_vapic_addr()
2053 vcpu->arch.apic->vapic_addr = vapic_addr; in kvm_lapic_set_vapic_addr()
2059 struct kvm_lapic *apic = vcpu->arch.apic; in kvm_x2apic_msr_write()
2076 struct kvm_lapic *apic = vcpu->arch.apic; in kvm_x2apic_msr_read()
2100 struct kvm_lapic *apic = vcpu->arch.apic; in kvm_hv_vapic_msr_write()
2113 struct kvm_lapic *apic = vcpu->arch.apic; in kvm_hv_vapic_msr_read()
2135 vcpu->arch.pv_eoi.msr_val = data; in kvm_lapic_enable_pv_eoi()
2138 return kvm_gfn_to_hva_cache_init(vcpu->kvm, &vcpu->arch.pv_eoi.data, in kvm_lapic_enable_pv_eoi()
2144 struct kvm_lapic *apic = vcpu->arch.apic; in kvm_apic_accept_events()
2157 WARN_ON_ONCE(vcpu->arch.mp_state == KVM_MP_STATE_INIT_RECEIVED); in kvm_apic_accept_events()
2168 vcpu->arch.mp_state = KVM_MP_STATE_RUNNABLE; in kvm_apic_accept_events()
2170 vcpu->arch.mp_state = KVM_MP_STATE_INIT_RECEIVED; in kvm_apic_accept_events()
2173 vcpu->arch.mp_state == KVM_MP_STATE_INIT_RECEIVED) { in kvm_apic_accept_events()
2180 vcpu->arch.mp_state = KVM_MP_STATE_RUNNABLE; in kvm_apic_accept_events()