Lines Matching refs:arch
51 return !!(v->arch.pending_exceptions) || in kvm_arch_vcpu_runnable()
131 struct kvm_vcpu_arch_shared *shared = vcpu->arch.shared; in kvmppc_swab_shared()
173 if (vcpu->arch.intr_msr & MSR_LE) in kvmppc_kvm_pv()
175 if (shared_big_endian != vcpu->arch.shared_big_endian) in kvmppc_kvm_pv()
177 vcpu->arch.shared_big_endian = shared_big_endian; in kvmppc_kvm_pv()
186 vcpu->arch.disable_kernel_nx = true; in kvmppc_kvm_pv()
190 vcpu->arch.magic_page_pa = param1 & ~0xfffULL; in kvmppc_kvm_pv()
191 vcpu->arch.magic_page_ea = param2 & ~0xfffULL; in kvmppc_kvm_pv()
198 if ((vcpu->arch.magic_page_pa & 0xf000) != in kvmppc_kvm_pv()
199 ((ulong)vcpu->arch.shared & 0xf000)) { in kvmppc_kvm_pv()
200 void *old_shared = vcpu->arch.shared; in kvmppc_kvm_pv()
201 ulong shared = (ulong)vcpu->arch.shared; in kvmppc_kvm_pv()
205 shared |= vcpu->arch.magic_page_pa & 0xf000; in kvmppc_kvm_pv()
208 vcpu->arch.shared = new_shared; in kvmppc_kvm_pv()
246 if (!vcpu->arch.pvr) in kvmppc_sanity_check()
250 if ((vcpu->arch.cpu_type != KVM_CPU_3S_64) && vcpu->arch.papr_enabled) in kvmppc_sanity_check()
254 if (!vcpu->arch.papr_enabled && is_kvmppc_hv_enabled(vcpu->kvm)) in kvmppc_sanity_check()
265 vcpu->arch.sane = r; in kvmppc_sanity_check()
315 ulong mp_pa = vcpu->arch.magic_page_pa & KVM_PAM & PAGE_MASK; in kvmppc_st()
335 void *magic = vcpu->arch.shared; in kvmppc_st()
351 ulong mp_pa = vcpu->arch.magic_page_pa & KVM_PAM & PAGE_MASK; in kvmppc_ld()
374 void *magic = vcpu->arch.shared; in kvmppc_ld()
429 kvm->arch.kvm_ops = kvm_ops; in kvm_arch_init_vm()
454 module_put(kvm->arch.kvm_ops->owner); in kvm_arch_destroy_vm()
627 vcpu->arch.wqp = &vcpu->wq; in kvm_arch_vcpu_create()
640 hrtimer_cancel(&vcpu->arch.dec_timer); in kvm_arch_vcpu_free()
644 switch (vcpu->arch.irq_type) { in kvm_arch_vcpu_free()
646 kvmppc_mpic_disconnect_vcpu(vcpu->arch.mpic, vcpu); in kvm_arch_vcpu_free()
670 vcpu = container_of(timer, struct kvm_vcpu, arch.dec_timer); in kvmppc_decrementer_wakeup()
680 hrtimer_init(&vcpu->arch.dec_timer, CLOCK_REALTIME, HRTIMER_MODE_ABS); in kvm_arch_vcpu_init()
681 vcpu->arch.dec_timer.function = kvmppc_decrementer_wakeup; in kvm_arch_vcpu_init()
682 vcpu->arch.dec_expires = ~(u64)0; in kvm_arch_vcpu_init()
685 mutex_init(&vcpu->arch.exit_timing_lock); in kvm_arch_vcpu_init()
707 mtspr(SPRN_VRSAVE, vcpu->arch.vrsave); in kvm_arch_vcpu_load()
716 vcpu->arch.vrsave = mfspr(SPRN_VRSAVE); in kvm_arch_vcpu_put()
730 if (!vcpu->arch.mmio_host_swabbed) { in kvmppc_complete_mmio_load()
746 if (vcpu->arch.mmio_sign_extend) { in kvmppc_complete_mmio_load()
762 kvmppc_set_gpr(vcpu, vcpu->arch.io_gpr, gpr); in kvmppc_complete_mmio_load()
764 switch (vcpu->arch.io_gpr & KVM_MMIO_REG_EXT_MASK) { in kvmppc_complete_mmio_load()
766 kvmppc_set_gpr(vcpu, vcpu->arch.io_gpr, gpr); in kvmppc_complete_mmio_load()
769 VCPU_FPR(vcpu, vcpu->arch.io_gpr & KVM_MMIO_REG_MASK) = gpr; in kvmppc_complete_mmio_load()
773 vcpu->arch.qpr[vcpu->arch.io_gpr & KVM_MMIO_REG_MASK] = gpr; in kvmppc_complete_mmio_load()
776 VCPU_FPR(vcpu, vcpu->arch.io_gpr & KVM_MMIO_REG_MASK) = gpr; in kvmppc_complete_mmio_load()
777 vcpu->arch.qpr[vcpu->arch.io_gpr & KVM_MMIO_REG_MASK] = gpr; in kvmppc_complete_mmio_load()
804 run->mmio.phys_addr = vcpu->arch.paddr_accessed; in kvmppc_handle_load()
808 vcpu->arch.io_gpr = rt; in kvmppc_handle_load()
809 vcpu->arch.mmio_host_swabbed = host_swabbed; in kvmppc_handle_load()
812 vcpu->arch.mmio_sign_extend = 0; in kvmppc_handle_load()
838 vcpu->arch.mmio_sign_extend = 1; in kvmppc_handle_loads()
863 run->mmio.phys_addr = vcpu->arch.paddr_accessed; in kvmppc_handle_store()
922 val.vval = vcpu->arch.vr.vr[reg->id - KVM_REG_PPC_VR0]; in kvm_vcpu_ioctl_get_one_reg()
929 val = get_reg_val(reg->id, vcpu->arch.vr.vscr.u[3]); in kvm_vcpu_ioctl_get_one_reg()
932 val = get_reg_val(reg->id, vcpu->arch.vrsave); in kvm_vcpu_ioctl_get_one_reg()
973 vcpu->arch.vr.vr[reg->id - KVM_REG_PPC_VR0] = val.vval; in kvm_vcpu_ioctl_set_one_reg()
980 vcpu->arch.vr.vscr.u[3] = set_reg_val(reg->id, val); in kvm_vcpu_ioctl_set_one_reg()
987 vcpu->arch.vrsave = set_reg_val(reg->id, val); in kvm_vcpu_ioctl_set_one_reg()
1011 } else if (vcpu->arch.osi_needed) { in kvm_arch_vcpu_ioctl_run()
1017 vcpu->arch.osi_needed = 0; in kvm_arch_vcpu_ioctl_run()
1018 } else if (vcpu->arch.hcall_needed) { in kvm_arch_vcpu_ioctl_run()
1024 vcpu->arch.hcall_needed = 0; in kvm_arch_vcpu_ioctl_run()
1026 } else if (vcpu->arch.epr_needed) { in kvm_arch_vcpu_ioctl_run()
1028 vcpu->arch.epr_needed = 0; in kvm_arch_vcpu_ioctl_run()
1065 vcpu->arch.osi_enabled = true; in kvm_vcpu_ioctl_enable_cap()
1069 vcpu->arch.papr_enabled = true; in kvm_vcpu_ioctl_enable_cap()
1074 vcpu->arch.epr_flags |= KVMPPC_EPR_USER; in kvm_vcpu_ioctl_enable_cap()
1076 vcpu->arch.epr_flags &= ~KVMPPC_EPR_USER; in kvm_vcpu_ioctl_enable_cap()
1081 vcpu->arch.watchdog_enabled = true; in kvm_vcpu_ioctl_enable_cap()
1290 set_bit(hcall / 4, kvm->arch.enabled_hcalls); in kvm_vm_ioctl_enable_cap()
1292 clear_bit(hcall / 4, kvm->arch.enabled_hcalls); in kvm_vm_ioctl_enable_cap()
1348 r = kvm->arch.kvm_ops->get_smmu_info(kvm, &info); in kvm_arch_vm_ioctl()
1361 r = kvm->arch.kvm_ops->arch_vm_ioctl(filp, ioctl, arg); in kvm_arch_vm_ioctl()