Lines Matching refs:kvm

254 	if (!vcpu->arch.papr_enabled && is_kvmppc_hv_enabled(vcpu->kvm))  in kvmppc_sanity_check()
341 if (kvm_write_guest(vcpu->kvm, pte.raddr, ptr, size)) in kvmppc_st()
380 if (kvm_read_guest(vcpu->kvm, pte.raddr, ptr, size)) in kvmppc_ld()
402 int kvm_arch_init_vm(struct kvm *kvm, unsigned long type) in kvm_arch_init_vm() argument
429 kvm->arch.kvm_ops = kvm_ops; in kvm_arch_init_vm()
430 return kvmppc_core_init_vm(kvm); in kvm_arch_init_vm()
435 void kvm_arch_destroy_vm(struct kvm *kvm) in kvm_arch_destroy_vm() argument
440 kvm_for_each_vcpu(i, vcpu, kvm) in kvm_arch_destroy_vm()
443 mutex_lock(&kvm->lock); in kvm_arch_destroy_vm()
444 for (i = 0; i < atomic_read(&kvm->online_vcpus); i++) in kvm_arch_destroy_vm()
445 kvm->vcpus[i] = NULL; in kvm_arch_destroy_vm()
447 atomic_set(&kvm->online_vcpus, 0); in kvm_arch_destroy_vm()
449 kvmppc_core_destroy_vm(kvm); in kvm_arch_destroy_vm()
451 mutex_unlock(&kvm->lock); in kvm_arch_destroy_vm()
454 module_put(kvm->arch.kvm_ops->owner); in kvm_arch_destroy_vm()
457 int kvm_vm_ioctl_check_extension(struct kvm *kvm, long ext) in kvm_vm_ioctl_check_extension() argument
463 if (kvm) { in kvm_vm_ioctl_check_extension()
468 hv_enabled = is_kvmppc_hv_enabled(kvm); in kvm_vm_ioctl_check_extension()
587 void kvm_arch_free_memslot(struct kvm *kvm, struct kvm_memory_slot *free, in kvm_arch_free_memslot() argument
590 kvmppc_core_free_memslot(kvm, free, dont); in kvm_arch_free_memslot()
593 int kvm_arch_create_memslot(struct kvm *kvm, struct kvm_memory_slot *slot, in kvm_arch_create_memslot() argument
596 return kvmppc_core_create_memslot(kvm, slot, npages); in kvm_arch_create_memslot()
599 int kvm_arch_prepare_memory_region(struct kvm *kvm, in kvm_arch_prepare_memory_region() argument
604 return kvmppc_core_prepare_memory_region(kvm, memslot, mem); in kvm_arch_prepare_memory_region()
607 void kvm_arch_commit_memory_region(struct kvm *kvm, in kvm_arch_commit_memory_region() argument
613 kvmppc_core_commit_memory_region(kvm, mem, old, new); in kvm_arch_commit_memory_region()
616 void kvm_arch_flush_shadow_memslot(struct kvm *kvm, in kvm_arch_flush_shadow_memslot() argument
619 kvmppc_core_flush_memslot(kvm, slot); in kvm_arch_flush_shadow_memslot()
622 struct kvm_vcpu *kvm_arch_vcpu_create(struct kvm *kvm, unsigned int id) in kvm_arch_vcpu_create() argument
625 vcpu = kvmppc_core_vcpu_create(kvm, id); in kvm_arch_vcpu_create()
814 idx = srcu_read_lock(&vcpu->kvm->srcu); in kvmppc_handle_load()
819 srcu_read_unlock(&vcpu->kvm->srcu, idx); in kvmppc_handle_load()
886 idx = srcu_read_lock(&vcpu->kvm->srcu); in kvmppc_handle_store()
891 srcu_read_unlock(&vcpu->kvm->srcu, idx); in kvmppc_handle_store()
1257 int kvm_vm_ioctl_irq_line(struct kvm *kvm, struct kvm_irq_level *irq_event, in kvm_vm_ioctl_irq_line() argument
1260 if (!irqchip_in_kernel(kvm)) in kvm_vm_ioctl_irq_line()
1263 irq_event->status = kvm_set_irq(kvm, KVM_USERSPACE_IRQ_SOURCE_ID, in kvm_vm_ioctl_irq_line()
1270 static int kvm_vm_ioctl_enable_cap(struct kvm *kvm, in kvm_vm_ioctl_enable_cap() argument
1287 if (!kvmppc_book3s_hcall_implemented(kvm, hcall)) in kvm_vm_ioctl_enable_cap()
1290 set_bit(hcall / 4, kvm->arch.enabled_hcalls); in kvm_vm_ioctl_enable_cap()
1292 clear_bit(hcall / 4, kvm->arch.enabled_hcalls); in kvm_vm_ioctl_enable_cap()
1308 struct kvm *kvm __maybe_unused = filp->private_data; in kvm_arch_vm_ioctl()
1330 r = kvm_vm_ioctl_enable_cap(kvm, &cap); in kvm_arch_vm_ioctl()
1340 r = kvm_vm_ioctl_create_spapr_tce(kvm, &create_tce); in kvm_arch_vm_ioctl()
1345 struct kvm *kvm = filp->private_data; in kvm_arch_vm_ioctl() local
1348 r = kvm->arch.kvm_ops->get_smmu_info(kvm, &info); in kvm_arch_vm_ioctl()
1354 struct kvm *kvm = filp->private_data; in kvm_arch_vm_ioctl() local
1356 r = kvm_vm_ioctl_rtas_define_token(kvm, argp); in kvm_arch_vm_ioctl()
1360 struct kvm *kvm = filp->private_data; in kvm_arch_vm_ioctl() local
1361 r = kvm->arch.kvm_ops->arch_vm_ioctl(filp, ioctl, arg); in kvm_arch_vm_ioctl()