Lines Matching refs:kvm
45 debug_sprintf_event(d_vcpu->kvm->arch.dbf, d_loglevel, \
56 static inline int kvm_is_ucontrol(struct kvm *kvm) in kvm_is_ucontrol() argument
59 if (kvm->arch.gmap) in kvm_is_ucontrol()
155 static inline int test_kvm_facility(struct kvm *kvm, unsigned long nr) in test_kvm_facility() argument
157 return __test_facility(nr, kvm->arch.model.fac->mask) && in test_kvm_facility()
158 __test_facility(nr, kvm->arch.model.fac->list); in test_kvm_facility()
173 static inline int kvm_s390_user_cpu_state_ctrl(struct kvm *kvm) in kvm_s390_user_cpu_state_ctrl() argument
175 return kvm->arch.user_cpu_state_ctrl != 0; in kvm_s390_user_cpu_state_ctrl()
184 void kvm_s390_clear_float_irqs(struct kvm *kvm);
185 int __must_check kvm_s390_inject_vm(struct kvm *kvm,
208 struct kvm_s390_interrupt_info *kvm_s390_get_io_int(struct kvm *kvm,
210 int kvm_s390_reinject_io_int(struct kvm *kvm,
212 int kvm_s390_mask_adapter(struct kvm *kvm, unsigned int id, bool masked);
234 void kvm_s390_set_tod_clock(struct kvm *kvm, u64 tod);
255 static inline void kvm_s390_vcpu_block_all(struct kvm *kvm) in kvm_s390_vcpu_block_all() argument
260 WARN_ON(!mutex_is_locked(&kvm->lock)); in kvm_s390_vcpu_block_all()
261 kvm_for_each_vcpu(i, vcpu, kvm) in kvm_s390_vcpu_block_all()
265 static inline void kvm_s390_vcpu_unblock_all(struct kvm *kvm) in kvm_s390_vcpu_unblock_all() argument
270 kvm_for_each_vcpu(i, vcpu, kvm) in kvm_s390_vcpu_unblock_all()
274 static inline u64 kvm_s390_get_tod_clock_fast(struct kvm *kvm) in kvm_s390_get_tod_clock_fast() argument
279 rc = get_tod_clock_fast() + kvm->arch.epoch; in kvm_s390_get_tod_clock_fast()
323 void kvm_s390_destroy_adapters(struct kvm *kvm);