Lines Matching refs:kvm
254 if (!vcpu->arch.papr_enabled && is_kvmppc_hv_enabled(vcpu->kvm)) in kvmppc_sanity_check()
341 if (kvm_write_guest(vcpu->kvm, pte.raddr, ptr, size)) in kvmppc_st()
380 if (kvm_read_guest(vcpu->kvm, pte.raddr, ptr, size)) in kvmppc_ld()
402 int kvm_arch_init_vm(struct kvm *kvm, unsigned long type) in kvm_arch_init_vm() argument
429 kvm->arch.kvm_ops = kvm_ops; in kvm_arch_init_vm()
430 return kvmppc_core_init_vm(kvm); in kvm_arch_init_vm()
435 void kvm_arch_destroy_vm(struct kvm *kvm) in kvm_arch_destroy_vm() argument
440 kvm_for_each_vcpu(i, vcpu, kvm) in kvm_arch_destroy_vm()
443 mutex_lock(&kvm->lock); in kvm_arch_destroy_vm()
444 for (i = 0; i < atomic_read(&kvm->online_vcpus); i++) in kvm_arch_destroy_vm()
445 kvm->vcpus[i] = NULL; in kvm_arch_destroy_vm()
447 atomic_set(&kvm->online_vcpus, 0); in kvm_arch_destroy_vm()
449 kvmppc_core_destroy_vm(kvm); in kvm_arch_destroy_vm()
451 mutex_unlock(&kvm->lock); in kvm_arch_destroy_vm()
454 module_put(kvm->arch.kvm_ops->owner); in kvm_arch_destroy_vm()
457 int kvm_vm_ioctl_check_extension(struct kvm *kvm, long ext) in kvm_vm_ioctl_check_extension() argument
463 if (kvm) { in kvm_vm_ioctl_check_extension()
468 hv_enabled = is_kvmppc_hv_enabled(kvm); in kvm_vm_ioctl_check_extension()
584 void kvm_arch_free_memslot(struct kvm *kvm, struct kvm_memory_slot *free, in kvm_arch_free_memslot() argument
587 kvmppc_core_free_memslot(kvm, free, dont); in kvm_arch_free_memslot()
590 int kvm_arch_create_memslot(struct kvm *kvm, struct kvm_memory_slot *slot, in kvm_arch_create_memslot() argument
593 return kvmppc_core_create_memslot(kvm, slot, npages); in kvm_arch_create_memslot()
596 int kvm_arch_prepare_memory_region(struct kvm *kvm, in kvm_arch_prepare_memory_region() argument
601 return kvmppc_core_prepare_memory_region(kvm, memslot, mem); in kvm_arch_prepare_memory_region()
604 void kvm_arch_commit_memory_region(struct kvm *kvm, in kvm_arch_commit_memory_region() argument
609 kvmppc_core_commit_memory_region(kvm, mem, old); in kvm_arch_commit_memory_region()
612 void kvm_arch_flush_shadow_memslot(struct kvm *kvm, in kvm_arch_flush_shadow_memslot() argument
615 kvmppc_core_flush_memslot(kvm, slot); in kvm_arch_flush_shadow_memslot()
618 struct kvm_vcpu *kvm_arch_vcpu_create(struct kvm *kvm, unsigned int id) in kvm_arch_vcpu_create() argument
621 vcpu = kvmppc_core_vcpu_create(kvm, id); in kvm_arch_vcpu_create()
810 idx = srcu_read_lock(&vcpu->kvm->srcu); in kvmppc_handle_load()
815 srcu_read_unlock(&vcpu->kvm->srcu, idx); in kvmppc_handle_load()
882 idx = srcu_read_lock(&vcpu->kvm->srcu); in kvmppc_handle_store()
887 srcu_read_unlock(&vcpu->kvm->srcu, idx); in kvmppc_handle_store()
1253 int kvm_vm_ioctl_irq_line(struct kvm *kvm, struct kvm_irq_level *irq_event, in kvm_vm_ioctl_irq_line() argument
1256 if (!irqchip_in_kernel(kvm)) in kvm_vm_ioctl_irq_line()
1259 irq_event->status = kvm_set_irq(kvm, KVM_USERSPACE_IRQ_SOURCE_ID, in kvm_vm_ioctl_irq_line()
1266 static int kvm_vm_ioctl_enable_cap(struct kvm *kvm, in kvm_vm_ioctl_enable_cap() argument
1283 if (!kvmppc_book3s_hcall_implemented(kvm, hcall)) in kvm_vm_ioctl_enable_cap()
1286 set_bit(hcall / 4, kvm->arch.enabled_hcalls); in kvm_vm_ioctl_enable_cap()
1288 clear_bit(hcall / 4, kvm->arch.enabled_hcalls); in kvm_vm_ioctl_enable_cap()
1304 struct kvm *kvm __maybe_unused = filp->private_data; in kvm_arch_vm_ioctl()
1326 r = kvm_vm_ioctl_enable_cap(kvm, &cap); in kvm_arch_vm_ioctl()
1336 r = kvm_vm_ioctl_create_spapr_tce(kvm, &create_tce); in kvm_arch_vm_ioctl()
1341 struct kvm *kvm = filp->private_data; in kvm_arch_vm_ioctl() local
1344 r = kvm->arch.kvm_ops->get_smmu_info(kvm, &info); in kvm_arch_vm_ioctl()
1350 struct kvm *kvm = filp->private_data; in kvm_arch_vm_ioctl() local
1352 r = kvm_vm_ioctl_rtas_define_token(kvm, argp); in kvm_arch_vm_ioctl()
1356 struct kvm *kvm = filp->private_data; in kvm_arch_vm_ioctl() local
1357 r = kvm->arch.kvm_ops->arch_vm_ioctl(filp, ioctl, arg); in kvm_arch_vm_ioctl()