kvm_vcpu 51 arch/arm/include/asm/kvm_asm.h struct kvm_vcpu; kvm_vcpu 59 arch/arm/include/asm/kvm_asm.h extern void __kvm_tlb_flush_local_vmid(struct kvm_vcpu *vcpu); kvm_vcpu 64 arch/arm/include/asm/kvm_asm.h static inline int kvm_vcpu_run_vhe(struct kvm_vcpu *vcpu) { BUG(); return 0; } kvm_vcpu 66 arch/arm/include/asm/kvm_asm.h extern int __kvm_vcpu_run_nvhe(struct kvm_vcpu *vcpu); kvm_vcpu 10 arch/arm/include/asm/kvm_coproc.h void kvm_reset_coprocs(struct kvm_vcpu *vcpu); kvm_vcpu 19 arch/arm/include/asm/kvm_coproc.h int kvm_handle_cp10_id(struct kvm_vcpu *vcpu, struct kvm_run *run); kvm_vcpu 20 arch/arm/include/asm/kvm_coproc.h int kvm_handle_cp_0_13_access(struct kvm_vcpu *vcpu, struct kvm_run *run); kvm_vcpu 21 arch/arm/include/asm/kvm_coproc.h int kvm_handle_cp14_load_store(struct kvm_vcpu *vcpu, struct kvm_run *run); kvm_vcpu 22 arch/arm/include/asm/kvm_coproc.h int kvm_handle_cp14_32(struct kvm_vcpu *vcpu, struct kvm_run *run); kvm_vcpu 23 arch/arm/include/asm/kvm_coproc.h int kvm_handle_cp14_64(struct kvm_vcpu *vcpu, struct kvm_run *run); kvm_vcpu 24 arch/arm/include/asm/kvm_coproc.h int kvm_handle_cp15_32(struct kvm_vcpu *vcpu, struct kvm_run *run); kvm_vcpu 25 arch/arm/include/asm/kvm_coproc.h int kvm_handle_cp15_64(struct kvm_vcpu *vcpu, struct kvm_run *run); kvm_vcpu 27 arch/arm/include/asm/kvm_coproc.h unsigned long kvm_arm_num_guest_msrs(struct kvm_vcpu *vcpu); kvm_vcpu 28 arch/arm/include/asm/kvm_coproc.h int kvm_arm_copy_msrindices(struct kvm_vcpu *vcpu, u64 __user *uindices); kvm_vcpu 32 arch/arm/include/asm/kvm_coproc.h int kvm_arm_copy_coproc_indices(struct kvm_vcpu *vcpu, u64 __user *uindices); kvm_vcpu 33 arch/arm/include/asm/kvm_coproc.h int kvm_arm_coproc_get_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *); kvm_vcpu 34 arch/arm/include/asm/kvm_coproc.h int kvm_arm_coproc_set_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *); kvm_vcpu 35 arch/arm/include/asm/kvm_coproc.h unsigned long kvm_arm_num_coproc_regs(struct kvm_vcpu *vcpu); kvm_vcpu 37 arch/arm/include/asm/kvm_emulate.h unsigned long *vcpu_reg(struct kvm_vcpu *vcpu, u8 reg_num); kvm_vcpu 39 arch/arm/include/asm/kvm_emulate.h static inline unsigned long *vcpu_reg32(struct kvm_vcpu *vcpu, u8 reg_num) kvm_vcpu 44 arch/arm/include/asm/kvm_emulate.h unsigned long *__vcpu_spsr(struct kvm_vcpu *vcpu); kvm_vcpu 46 arch/arm/include/asm/kvm_emulate.h static inline unsigned long vpcu_read_spsr(struct kvm_vcpu *vcpu) kvm_vcpu 51 arch/arm/include/asm/kvm_emulate.h static inline void vcpu_write_spsr(struct kvm_vcpu *vcpu, unsigned long v) kvm_vcpu 61 arch/arm/include/asm/kvm_emulate.h static inline unsigned long vcpu_get_reg(struct kvm_vcpu *vcpu, kvm_vcpu 67 arch/arm/include/asm/kvm_emulate.h static inline void vcpu_set_reg(struct kvm_vcpu *vcpu, u8 reg_num, kvm_vcpu 73 arch/arm/include/asm/kvm_emulate.h bool kvm_condition_valid32(const struct kvm_vcpu *vcpu); kvm_vcpu 74 arch/arm/include/asm/kvm_emulate.h void kvm_skip_instr32(struct kvm_vcpu *vcpu, bool is_wide_instr); kvm_vcpu 75 arch/arm/include/asm/kvm_emulate.h void kvm_inject_undef32(struct kvm_vcpu *vcpu); kvm_vcpu 76 arch/arm/include/asm/kvm_emulate.h void kvm_inject_dabt32(struct kvm_vcpu *vcpu, unsigned long addr); kvm_vcpu 77 arch/arm/include/asm/kvm_emulate.h void kvm_inject_pabt32(struct kvm_vcpu *vcpu, unsigned long addr); kvm_vcpu 78 arch/arm/include/asm/kvm_emulate.h void kvm_inject_vabt(struct kvm_vcpu *vcpu); kvm_vcpu 80 arch/arm/include/asm/kvm_emulate.h static inline void kvm_inject_undefined(struct kvm_vcpu *vcpu) kvm_vcpu 85 arch/arm/include/asm/kvm_emulate.h static inline void kvm_inject_dabt(struct kvm_vcpu *vcpu, unsigned long addr) kvm_vcpu 90 arch/arm/include/asm/kvm_emulate.h static inline void kvm_inject_pabt(struct kvm_vcpu *vcpu, unsigned long addr) kvm_vcpu 95 arch/arm/include/asm/kvm_emulate.h static inline bool kvm_condition_valid(const struct kvm_vcpu *vcpu) kvm_vcpu 100 arch/arm/include/asm/kvm_emulate.h static inline void kvm_skip_instr(struct kvm_vcpu *vcpu, bool is_wide_instr) kvm_vcpu 105 arch/arm/include/asm/kvm_emulate.h static inline void vcpu_reset_hcr(struct kvm_vcpu *vcpu) kvm_vcpu 110 arch/arm/include/asm/kvm_emulate.h static inline unsigned long *vcpu_hcr(const struct kvm_vcpu *vcpu) kvm_vcpu 115 arch/arm/include/asm/kvm_emulate.h static inline void vcpu_clear_wfe_traps(struct kvm_vcpu *vcpu) kvm_vcpu 120 arch/arm/include/asm/kvm_emulate.h static inline void vcpu_set_wfe_traps(struct kvm_vcpu *vcpu) kvm_vcpu 125 arch/arm/include/asm/kvm_emulate.h static inline bool vcpu_mode_is_32bit(const struct kvm_vcpu *vcpu) kvm_vcpu 130 arch/arm/include/asm/kvm_emulate.h static inline unsigned long *vcpu_pc(struct kvm_vcpu *vcpu) kvm_vcpu 135 arch/arm/include/asm/kvm_emulate.h static inline unsigned long *vcpu_cpsr(const struct kvm_vcpu *vcpu) kvm_vcpu 140 arch/arm/include/asm/kvm_emulate.h static inline void vcpu_set_thumb(struct kvm_vcpu *vcpu) kvm_vcpu 145 arch/arm/include/asm/kvm_emulate.h static inline bool mode_has_spsr(struct kvm_vcpu *vcpu) kvm_vcpu 151 arch/arm/include/asm/kvm_emulate.h static inline bool vcpu_mode_priv(struct kvm_vcpu *vcpu) kvm_vcpu 157 arch/arm/include/asm/kvm_emulate.h static inline u32 kvm_vcpu_get_hsr(const struct kvm_vcpu *vcpu) kvm_vcpu 162 arch/arm/include/asm/kvm_emulate.h static inline int kvm_vcpu_get_condition(const struct kvm_vcpu *vcpu) kvm_vcpu 172 arch/arm/include/asm/kvm_emulate.h static inline unsigned long kvm_vcpu_get_hfar(struct kvm_vcpu *vcpu) kvm_vcpu 177 arch/arm/include/asm/kvm_emulate.h static inline phys_addr_t kvm_vcpu_get_fault_ipa(struct kvm_vcpu *vcpu) kvm_vcpu 182 arch/arm/include/asm/kvm_emulate.h static inline bool kvm_vcpu_dabt_isvalid(struct kvm_vcpu *vcpu) kvm_vcpu 187 arch/arm/include/asm/kvm_emulate.h static inline bool kvm_vcpu_dabt_iswrite(struct kvm_vcpu *vcpu) kvm_vcpu 192 arch/arm/include/asm/kvm_emulate.h static inline bool kvm_vcpu_dabt_issext(struct kvm_vcpu *vcpu) kvm_vcpu 197 arch/arm/include/asm/kvm_emulate.h static inline bool kvm_vcpu_dabt_issf(const struct kvm_vcpu *vcpu) kvm_vcpu 202 arch/arm/include/asm/kvm_emulate.h static inline int kvm_vcpu_dabt_get_rd(struct kvm_vcpu *vcpu) kvm_vcpu 207 arch/arm/include/asm/kvm_emulate.h static inline bool kvm_vcpu_dabt_iss1tw(struct kvm_vcpu *vcpu) kvm_vcpu 212 arch/arm/include/asm/kvm_emulate.h static inline bool kvm_vcpu_dabt_is_cm(struct kvm_vcpu *vcpu) kvm_vcpu 218 arch/arm/include/asm/kvm_emulate.h static inline int kvm_vcpu_dabt_get_as(struct kvm_vcpu *vcpu) kvm_vcpu 234 arch/arm/include/asm/kvm_emulate.h static inline bool kvm_vcpu_trap_il_is32bit(struct kvm_vcpu *vcpu) kvm_vcpu 239 arch/arm/include/asm/kvm_emulate.h static inline u8 kvm_vcpu_trap_get_class(struct kvm_vcpu *vcpu) kvm_vcpu 244 arch/arm/include/asm/kvm_emulate.h static inline bool kvm_vcpu_trap_is_iabt(struct kvm_vcpu *vcpu) kvm_vcpu 249 arch/arm/include/asm/kvm_emulate.h static inline u8 kvm_vcpu_trap_get_fault(struct kvm_vcpu *vcpu) kvm_vcpu 254 arch/arm/include/asm/kvm_emulate.h static inline u8 kvm_vcpu_trap_get_fault_type(struct kvm_vcpu *vcpu) kvm_vcpu 259 arch/arm/include/asm/kvm_emulate.h static inline bool kvm_vcpu_dabt_isextabt(struct kvm_vcpu *vcpu) kvm_vcpu 278 arch/arm/include/asm/kvm_emulate.h static inline bool kvm_is_write_fault(struct kvm_vcpu *vcpu) kvm_vcpu 286 arch/arm/include/asm/kvm_emulate.h static inline u32 kvm_vcpu_hvc_get_imm(struct kvm_vcpu *vcpu) kvm_vcpu 291 arch/arm/include/asm/kvm_emulate.h static inline unsigned long kvm_vcpu_get_mpidr_aff(struct kvm_vcpu *vcpu) kvm_vcpu 296 arch/arm/include/asm/kvm_emulate.h static inline bool kvm_arm_get_vcpu_workaround_2_flag(struct kvm_vcpu *vcpu) kvm_vcpu 301 arch/arm/include/asm/kvm_emulate.h static inline void kvm_arm_set_vcpu_workaround_2_flag(struct kvm_vcpu *vcpu, kvm_vcpu 306 arch/arm/include/asm/kvm_emulate.h static inline void kvm_vcpu_set_be(struct kvm_vcpu *vcpu) kvm_vcpu 311 arch/arm/include/asm/kvm_emulate.h static inline bool kvm_vcpu_is_be(struct kvm_vcpu *vcpu) kvm_vcpu 316 arch/arm/include/asm/kvm_emulate.h static inline unsigned long vcpu_data_guest_to_host(struct kvm_vcpu *vcpu, kvm_vcpu 341 arch/arm/include/asm/kvm_emulate.h static inline unsigned long vcpu_data_host_to_guest(struct kvm_vcpu *vcpu, kvm_vcpu 366 arch/arm/include/asm/kvm_emulate.h static inline bool vcpu_has_ptrauth(struct kvm_vcpu *vcpu) { return false; } kvm_vcpu 367 arch/arm/include/asm/kvm_emulate.h static inline void vcpu_ptrauth_disable(struct kvm_vcpu *vcpu) { } kvm_vcpu 46 arch/arm/include/asm/kvm_host.h u32 *kvm_vcpu_reg(struct kvm_vcpu *vcpu, u8 reg_num, u32 mode); kvm_vcpu 48 arch/arm/include/asm/kvm_host.h int kvm_reset_vcpu(struct kvm_vcpu *vcpu); kvm_vcpu 49 arch/arm/include/asm/kvm_host.h void kvm_reset_coprocs(struct kvm_vcpu *vcpu); kvm_vcpu 227 arch/arm/include/asm/kvm_host.h unsigned long kvm_arm_num_regs(struct kvm_vcpu *vcpu); kvm_vcpu 228 arch/arm/include/asm/kvm_host.h int kvm_arm_copy_reg_indices(struct kvm_vcpu *vcpu, u64 __user *indices); kvm_vcpu 229 arch/arm/include/asm/kvm_host.h int kvm_arm_get_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg); kvm_vcpu 230 arch/arm/include/asm/kvm_host.h int kvm_arm_set_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg); kvm_vcpu 261 arch/arm/include/asm/kvm_host.h int __kvm_arm_vcpu_get_events(struct kvm_vcpu *vcpu, kvm_vcpu 264 arch/arm/include/asm/kvm_host.h int __kvm_arm_vcpu_set_events(struct kvm_vcpu *vcpu, kvm_vcpu 272 arch/arm/include/asm/kvm_host.h unsigned long kvm_arm_num_regs(struct kvm_vcpu *vcpu); kvm_vcpu 273 arch/arm/include/asm/kvm_host.h int kvm_arm_copy_reg_indices(struct kvm_vcpu *vcpu, u64 __user *indices); kvm_vcpu 277 arch/arm/include/asm/kvm_host.h struct kvm_vcpu *kvm_arm_get_running_vcpu(void); kvm_vcpu 278 arch/arm/include/asm/kvm_host.h struct kvm_vcpu __percpu **kvm_get_running_vcpus(void); kvm_vcpu 282 arch/arm/include/asm/kvm_host.h int kvm_arm_copy_coproc_indices(struct kvm_vcpu *vcpu, u64 __user *uindices); kvm_vcpu 283 arch/arm/include/asm/kvm_host.h unsigned long kvm_arm_num_coproc_regs(struct kvm_vcpu *vcpu); kvm_vcpu 284 arch/arm/include/asm/kvm_host.h int kvm_arm_coproc_get_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *); kvm_vcpu 285 arch/arm/include/asm/kvm_host.h int kvm_arm_coproc_set_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *); kvm_vcpu 287 arch/arm/include/asm/kvm_host.h int handle_exit(struct kvm_vcpu *vcpu, struct kvm_run *run, kvm_vcpu 290 arch/arm/include/asm/kvm_host.h static inline void handle_exit_early(struct kvm_vcpu *vcpu, struct kvm_run *run, kvm_vcpu 328 arch/arm/include/asm/kvm_host.h struct kvm_vcpu *kvm_mpidr_to_vcpu(struct kvm *kvm, unsigned long mpidr); kvm_vcpu 333 arch/arm/include/asm/kvm_host.h static inline void kvm_arch_vcpu_uninit(struct kvm_vcpu *vcpu) {} kvm_vcpu 334 arch/arm/include/asm/kvm_host.h static inline void kvm_arch_sched_in(struct kvm_vcpu *vcpu, int cpu) {} kvm_vcpu 335 arch/arm/include/asm/kvm_host.h static inline void kvm_arch_vcpu_block_finish(struct kvm_vcpu *vcpu) {} kvm_vcpu 338 arch/arm/include/asm/kvm_host.h static inline void kvm_arm_setup_debug(struct kvm_vcpu *vcpu) {} kvm_vcpu 339 arch/arm/include/asm/kvm_host.h static inline void kvm_arm_clear_debug(struct kvm_vcpu *vcpu) {} kvm_vcpu 340 arch/arm/include/asm/kvm_host.h static inline void kvm_arm_reset_debug_ptr(struct kvm_vcpu *vcpu) {} kvm_vcpu 342 arch/arm/include/asm/kvm_host.h int kvm_arm_vcpu_arch_set_attr(struct kvm_vcpu *vcpu, kvm_vcpu 344 arch/arm/include/asm/kvm_host.h int kvm_arm_vcpu_arch_get_attr(struct kvm_vcpu *vcpu, kvm_vcpu 346 arch/arm/include/asm/kvm_host.h int kvm_arm_vcpu_arch_has_attr(struct kvm_vcpu *vcpu, kvm_vcpu 353 arch/arm/include/asm/kvm_host.h static inline void kvm_arch_vcpu_load_fp(struct kvm_vcpu *vcpu) {} kvm_vcpu 354 arch/arm/include/asm/kvm_host.h static inline void kvm_arch_vcpu_ctxsync_fp(struct kvm_vcpu *vcpu) {} kvm_vcpu 355 arch/arm/include/asm/kvm_host.h static inline void kvm_arch_vcpu_put_fp(struct kvm_vcpu *vcpu) {} kvm_vcpu 357 arch/arm/include/asm/kvm_host.h static inline void kvm_vcpu_pmu_restore_guest(struct kvm_vcpu *vcpu) {} kvm_vcpu 358 arch/arm/include/asm/kvm_host.h static inline void kvm_vcpu_pmu_restore_host(struct kvm_vcpu *vcpu) {} kvm_vcpu 396 arch/arm/include/asm/kvm_host.h static inline void kvm_vcpu_load_sysregs(struct kvm_vcpu *vcpu) {} kvm_vcpu 397 arch/arm/include/asm/kvm_host.h static inline void kvm_vcpu_put_sysregs(struct kvm_vcpu *vcpu) {} kvm_vcpu 414 arch/arm/include/asm/kvm_host.h static inline int kvm_arm_vcpu_finalize(struct kvm_vcpu *vcpu, int feature) kvm_vcpu 419 arch/arm/include/asm/kvm_host.h static inline bool kvm_arm_vcpu_is_finalized(struct kvm_vcpu *vcpu) kvm_vcpu 96 arch/arm/include/asm/kvm_hyp.h void __timer_enable_traps(struct kvm_vcpu *vcpu); kvm_vcpu 97 arch/arm/include/asm/kvm_hyp.h void __timer_disable_traps(struct kvm_vcpu *vcpu); kvm_vcpu 99 arch/arm/include/asm/kvm_hyp.h void __vgic_v2_save_state(struct kvm_vcpu *vcpu); kvm_vcpu 100 arch/arm/include/asm/kvm_hyp.h void __vgic_v2_restore_state(struct kvm_vcpu *vcpu); kvm_vcpu 105 arch/arm/include/asm/kvm_hyp.h void __vgic_v3_save_state(struct kvm_vcpu *vcpu); kvm_vcpu 106 arch/arm/include/asm/kvm_hyp.h void __vgic_v3_restore_state(struct kvm_vcpu *vcpu); kvm_vcpu 107 arch/arm/include/asm/kvm_hyp.h void __vgic_v3_activate_traps(struct kvm_vcpu *vcpu); kvm_vcpu 108 arch/arm/include/asm/kvm_hyp.h void __vgic_v3_deactivate_traps(struct kvm_vcpu *vcpu); kvm_vcpu 109 arch/arm/include/asm/kvm_hyp.h void __vgic_v3_save_aprs(struct kvm_vcpu *vcpu); kvm_vcpu 110 arch/arm/include/asm/kvm_hyp.h void __vgic_v3_restore_aprs(struct kvm_vcpu *vcpu); kvm_vcpu 122 arch/arm/include/asm/kvm_hyp.h asmlinkage int __guest_enter(struct kvm_vcpu *vcpu, kvm_vcpu 24 arch/arm/include/asm/kvm_mmio.h int kvm_handle_mmio_return(struct kvm_vcpu *vcpu, struct kvm_run *run); kvm_vcpu 25 arch/arm/include/asm/kvm_mmio.h int io_mem_abort(struct kvm_vcpu *vcpu, struct kvm_run *run, kvm_vcpu 60 arch/arm/include/asm/kvm_mmu.h int kvm_handle_guest_abort(struct kvm_vcpu *vcpu, struct kvm_run *run); kvm_vcpu 62 arch/arm/include/asm/kvm_mmu.h void kvm_mmu_free_memory_caches(struct kvm_vcpu *vcpu); kvm_vcpu 206 arch/arm/include/asm/kvm_mmu.h static inline bool vcpu_has_cache_enabled(struct kvm_vcpu *vcpu) kvm_vcpu 333 arch/arm/include/asm/kvm_mmu.h void kvm_set_way_flush(struct kvm_vcpu *vcpu); kvm_vcpu 334 arch/arm/include/asm/kvm_mmu.h void kvm_toggle_cache(struct kvm_vcpu *vcpu, bool was_enabled); kvm_vcpu 171 arch/arm/kernel/asm-offsets.c DEFINE(VCPU_GUEST_CTXT, offsetof(struct kvm_vcpu, arch.ctxt)); kvm_vcpu 172 arch/arm/kernel/asm-offsets.c DEFINE(VCPU_HOST_CTXT, offsetof(struct kvm_vcpu, arch.host_cpu_context)); kvm_vcpu 32 arch/arm/kvm/coproc.c static bool write_to_read_only(struct kvm_vcpu *vcpu, kvm_vcpu 41 arch/arm/kvm/coproc.c static bool read_from_write_only(struct kvm_vcpu *vcpu, kvm_vcpu 62 arch/arm/kvm/coproc.c static inline void vcpu_cp15_reg64_set(struct kvm_vcpu *vcpu, kvm_vcpu 70 arch/arm/kvm/coproc.c static inline u64 vcpu_cp15_reg64_get(struct kvm_vcpu *vcpu, kvm_vcpu 81 arch/arm/kvm/coproc.c int kvm_handle_cp10_id(struct kvm_vcpu *vcpu, struct kvm_run *run) kvm_vcpu 87 arch/arm/kvm/coproc.c int kvm_handle_cp_0_13_access(struct kvm_vcpu *vcpu, struct kvm_run *run) kvm_vcpu 97 arch/arm/kvm/coproc.c int kvm_handle_cp14_load_store(struct kvm_vcpu *vcpu, struct kvm_run *run) kvm_vcpu 103 arch/arm/kvm/coproc.c static void reset_mpidr(struct kvm_vcpu *vcpu, const struct coproc_reg *r) kvm_vcpu 116 arch/arm/kvm/coproc.c static bool access_actlr(struct kvm_vcpu *vcpu, kvm_vcpu 128 arch/arm/kvm/coproc.c static bool access_cbar(struct kvm_vcpu *vcpu, kvm_vcpu 138 arch/arm/kvm/coproc.c static bool access_l2ctlr(struct kvm_vcpu *vcpu, kvm_vcpu 149 arch/arm/kvm/coproc.c static void reset_l2ctlr(struct kvm_vcpu *vcpu, const struct coproc_reg *r) kvm_vcpu 165 arch/arm/kvm/coproc.c static void reset_actlr(struct kvm_vcpu *vcpu, const struct coproc_reg *r) kvm_vcpu 184 arch/arm/kvm/coproc.c static bool access_l2ectlr(struct kvm_vcpu *vcpu, kvm_vcpu 198 arch/arm/kvm/coproc.c static bool access_dcsw(struct kvm_vcpu *vcpu, kvm_vcpu 216 arch/arm/kvm/coproc.c bool access_vm_reg(struct kvm_vcpu *vcpu, kvm_vcpu 232 arch/arm/kvm/coproc.c static bool access_gic_sgi(struct kvm_vcpu *vcpu, kvm_vcpu 268 arch/arm/kvm/coproc.c static bool access_gic_sre(struct kvm_vcpu *vcpu, kvm_vcpu 280 arch/arm/kvm/coproc.c static bool access_cntp_tval(struct kvm_vcpu *vcpu, kvm_vcpu 299 arch/arm/kvm/coproc.c static bool access_cntp_ctl(struct kvm_vcpu *vcpu, kvm_vcpu 318 arch/arm/kvm/coproc.c static bool access_cntp_cval(struct kvm_vcpu *vcpu, kvm_vcpu 348 arch/arm/kvm/coproc.c static bool trap_raz_wi(struct kvm_vcpu *vcpu, kvm_vcpu 572 arch/arm/kvm/coproc.c static int emulate_cp15(struct kvm_vcpu *vcpu, kvm_vcpu 607 arch/arm/kvm/coproc.c static struct coproc_params decode_64bit_hsr(struct kvm_vcpu *vcpu) kvm_vcpu 629 arch/arm/kvm/coproc.c int kvm_handle_cp15_64(struct kvm_vcpu *vcpu, struct kvm_run *run) kvm_vcpu 641 arch/arm/kvm/coproc.c int kvm_handle_cp14_64(struct kvm_vcpu *vcpu, struct kvm_run *run) kvm_vcpu 653 arch/arm/kvm/coproc.c static void reset_coproc_regs(struct kvm_vcpu *vcpu, kvm_vcpu 672 arch/arm/kvm/coproc.c static struct coproc_params decode_32bit_hsr(struct kvm_vcpu *vcpu) kvm_vcpu 694 arch/arm/kvm/coproc.c int kvm_handle_cp15_32(struct kvm_vcpu *vcpu, struct kvm_run *run) kvm_vcpu 705 arch/arm/kvm/coproc.c int kvm_handle_cp14_32(struct kvm_vcpu *vcpu, struct kvm_run *run) kvm_vcpu 766 arch/arm/kvm/coproc.c static const struct coproc_reg *index_to_coproc_reg(struct kvm_vcpu *vcpu, kvm_vcpu 801 arch/arm/kvm/coproc.c static void get_##name(struct kvm_vcpu *v, \ kvm_vcpu 1086 arch/arm/kvm/coproc.c static int vfp_get_reg(const struct kvm_vcpu *vcpu, u64 id, void __user *uaddr) kvm_vcpu 1130 arch/arm/kvm/coproc.c static int vfp_set_reg(struct kvm_vcpu *vcpu, u64 id, const void __user *uaddr) kvm_vcpu 1194 arch/arm/kvm/coproc.c static int vfp_get_reg(const struct kvm_vcpu *vcpu, u64 id, void __user *uaddr) kvm_vcpu 1199 arch/arm/kvm/coproc.c static int vfp_set_reg(struct kvm_vcpu *vcpu, u64 id, const void __user *uaddr) kvm_vcpu 1205 arch/arm/kvm/coproc.c int kvm_arm_coproc_get_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg) kvm_vcpu 1234 arch/arm/kvm/coproc.c int kvm_arm_coproc_set_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg) kvm_vcpu 1328 arch/arm/kvm/coproc.c static int walk_cp15(struct kvm_vcpu *vcpu, u64 __user *uind) kvm_vcpu 1370 arch/arm/kvm/coproc.c unsigned long kvm_arm_num_coproc_regs(struct kvm_vcpu *vcpu) kvm_vcpu 1378 arch/arm/kvm/coproc.c int kvm_arm_copy_coproc_indices(struct kvm_vcpu *vcpu, u64 __user *uindices) kvm_vcpu 1440 arch/arm/kvm/coproc.c void kvm_reset_coprocs(struct kvm_vcpu *vcpu) kvm_vcpu 31 arch/arm/kvm/coproc.h bool (*access)(struct kvm_vcpu *, kvm_vcpu 36 arch/arm/kvm/coproc.h void (*reset)(struct kvm_vcpu *, const struct coproc_reg *); kvm_vcpu 59 arch/arm/kvm/coproc.h static inline bool ignore_write(struct kvm_vcpu *vcpu, kvm_vcpu 65 arch/arm/kvm/coproc.h static inline bool read_zero(struct kvm_vcpu *vcpu, kvm_vcpu 73 arch/arm/kvm/coproc.h static inline void reset_unknown(struct kvm_vcpu *vcpu, kvm_vcpu 81 arch/arm/kvm/coproc.h static inline void reset_val(struct kvm_vcpu *vcpu, const struct coproc_reg *r) kvm_vcpu 88 arch/arm/kvm/coproc.h static inline void reset_unknown64(struct kvm_vcpu *vcpu, kvm_vcpu 126 arch/arm/kvm/coproc.h bool access_vm_reg(struct kvm_vcpu *vcpu, kvm_vcpu 101 arch/arm/kvm/emulate.c unsigned long *vcpu_reg(struct kvm_vcpu *vcpu, u8 reg_num) kvm_vcpu 133 arch/arm/kvm/emulate.c unsigned long *__vcpu_spsr(struct kvm_vcpu *vcpu) kvm_vcpu 163 arch/arm/kvm/emulate.c void kvm_inject_vabt(struct kvm_vcpu *vcpu) kvm_vcpu 21 arch/arm/kvm/guest.c #define VCPU_STAT(x) { #x, offsetof(struct kvm_vcpu, stat.x), KVM_STAT_VCPU } kvm_vcpu 33 arch/arm/kvm/guest.c int kvm_arch_vcpu_setup(struct kvm_vcpu *vcpu) kvm_vcpu 43 arch/arm/kvm/guest.c static int get_core_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg) kvm_vcpu 60 arch/arm/kvm/guest.c static int set_core_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg) kvm_vcpu 96 arch/arm/kvm/guest.c int kvm_arch_vcpu_ioctl_get_regs(struct kvm_vcpu *vcpu, struct kvm_regs *regs) kvm_vcpu 101 arch/arm/kvm/guest.c int kvm_arch_vcpu_ioctl_set_regs(struct kvm_vcpu *vcpu, struct kvm_regs *regs) kvm_vcpu 119 arch/arm/kvm/guest.c static int copy_timer_indices(struct kvm_vcpu *vcpu, u64 __user *uindices) kvm_vcpu 133 arch/arm/kvm/guest.c static int set_timer_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg) kvm_vcpu 146 arch/arm/kvm/guest.c static int get_timer_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg) kvm_vcpu 165 arch/arm/kvm/guest.c unsigned long kvm_arm_num_regs(struct kvm_vcpu *vcpu) kvm_vcpu 177 arch/arm/kvm/guest.c int kvm_arm_copy_reg_indices(struct kvm_vcpu *vcpu, u64 __user *uindices) kvm_vcpu 202 arch/arm/kvm/guest.c int kvm_arm_get_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg) kvm_vcpu 221 arch/arm/kvm/guest.c int kvm_arm_set_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg) kvm_vcpu 240 arch/arm/kvm/guest.c int kvm_arch_vcpu_ioctl_get_sregs(struct kvm_vcpu *vcpu, kvm_vcpu 246 arch/arm/kvm/guest.c int kvm_arch_vcpu_ioctl_set_sregs(struct kvm_vcpu *vcpu, kvm_vcpu 253 arch/arm/kvm/guest.c int __kvm_arm_vcpu_get_events(struct kvm_vcpu *vcpu, kvm_vcpu 261 arch/arm/kvm/guest.c int __kvm_arm_vcpu_set_events(struct kvm_vcpu *vcpu, kvm_vcpu 307 arch/arm/kvm/guest.c int kvm_arch_vcpu_ioctl_get_fpu(struct kvm_vcpu *vcpu, struct kvm_fpu *fpu) kvm_vcpu 312 arch/arm/kvm/guest.c int kvm_arch_vcpu_ioctl_set_fpu(struct kvm_vcpu *vcpu, struct kvm_fpu *fpu) kvm_vcpu 317 arch/arm/kvm/guest.c int kvm_arch_vcpu_ioctl_translate(struct kvm_vcpu *vcpu, kvm_vcpu 323 arch/arm/kvm/guest.c int kvm_arch_vcpu_ioctl_set_guest_debug(struct kvm_vcpu *vcpu, kvm_vcpu 329 arch/arm/kvm/guest.c int kvm_arm_vcpu_arch_set_attr(struct kvm_vcpu *vcpu, kvm_vcpu 346 arch/arm/kvm/guest.c int kvm_arm_vcpu_arch_get_attr(struct kvm_vcpu *vcpu, kvm_vcpu 363 arch/arm/kvm/guest.c int kvm_arm_vcpu_arch_has_attr(struct kvm_vcpu *vcpu, kvm_vcpu 17 arch/arm/kvm/handle_exit.c typedef int (*exit_handle_fn)(struct kvm_vcpu *, struct kvm_run *); kvm_vcpu 19 arch/arm/kvm/handle_exit.c static int handle_hvc(struct kvm_vcpu *vcpu, struct kvm_run *run) kvm_vcpu 36 arch/arm/kvm/handle_exit.c static int handle_smc(struct kvm_vcpu *vcpu, struct kvm_run *run) kvm_vcpu 62 arch/arm/kvm/handle_exit.c static int kvm_handle_wfx(struct kvm_vcpu *vcpu, struct kvm_run *run) kvm_vcpu 80 arch/arm/kvm/handle_exit.c static int kvm_handle_unknown_ec(struct kvm_vcpu *vcpu, struct kvm_run *run) kvm_vcpu 107 arch/arm/kvm/handle_exit.c static exit_handle_fn kvm_get_exit_handler(struct kvm_vcpu *vcpu) kvm_vcpu 118 arch/arm/kvm/handle_exit.c int handle_exit(struct kvm_vcpu *vcpu, struct kvm_run *run, kvm_vcpu 18 arch/arm/kvm/hyp/switch.c static void __hyp_text __activate_traps(struct kvm_vcpu *vcpu, u32 *fpexc_host) kvm_vcpu 46 arch/arm/kvm/hyp/switch.c static void __hyp_text __deactivate_traps(struct kvm_vcpu *vcpu) kvm_vcpu 66 arch/arm/kvm/hyp/switch.c static void __hyp_text __activate_vm(struct kvm_vcpu *vcpu) kvm_vcpu 73 arch/arm/kvm/hyp/switch.c static void __hyp_text __deactivate_vm(struct kvm_vcpu *vcpu) kvm_vcpu 80 arch/arm/kvm/hyp/switch.c static void __hyp_text __vgic_save_state(struct kvm_vcpu *vcpu) kvm_vcpu 88 arch/arm/kvm/hyp/switch.c static void __hyp_text __vgic_restore_state(struct kvm_vcpu *vcpu) kvm_vcpu 96 arch/arm/kvm/hyp/switch.c static bool __hyp_text __populate_fault_info(struct kvm_vcpu *vcpu) kvm_vcpu 146 arch/arm/kvm/hyp/switch.c int __hyp_text __kvm_vcpu_run_nvhe(struct kvm_vcpu *vcpu) kvm_vcpu 226 arch/arm/kvm/hyp/switch.c struct kvm_vcpu *vcpu; kvm_vcpu 229 arch/arm/kvm/hyp/switch.c vcpu = (struct kvm_vcpu *)read_sysreg(HTPIDR); kvm_vcpu 48 arch/arm/kvm/hyp/tlb.c void __hyp_text __kvm_tlb_flush_local_vmid(struct kvm_vcpu *vcpu) kvm_vcpu 41 arch/arm/kvm/reset.c int kvm_reset_vcpu(struct kvm_vcpu *vcpu) kvm_vcpu 11 arch/arm/kvm/vgic-v3-coproc.c int vgic_v3_has_cpu_sysregs_attr(struct kvm_vcpu *vcpu, bool is_write, u64 id, kvm_vcpu 20 arch/arm/kvm/vgic-v3-coproc.c int vgic_v3_cpu_sysregs_uaccess(struct kvm_vcpu *vcpu, bool is_write, u64 id, kvm_vcpu 53 arch/arm64/include/asm/kvm_asm.h struct kvm_vcpu; kvm_vcpu 63 arch/arm64/include/asm/kvm_asm.h extern void __kvm_tlb_flush_local_vmid(struct kvm_vcpu *vcpu); kvm_vcpu 67 arch/arm64/include/asm/kvm_asm.h extern int kvm_vcpu_run_vhe(struct kvm_vcpu *vcpu); kvm_vcpu 69 arch/arm64/include/asm/kvm_asm.h extern int __kvm_vcpu_run_nvhe(struct kvm_vcpu *vcpu); kvm_vcpu 15 arch/arm64/include/asm/kvm_coproc.h void kvm_reset_sys_regs(struct kvm_vcpu *vcpu); kvm_vcpu 30 arch/arm64/include/asm/kvm_coproc.h int kvm_handle_cp14_load_store(struct kvm_vcpu *vcpu, struct kvm_run *run); kvm_vcpu 31 arch/arm64/include/asm/kvm_coproc.h int kvm_handle_cp14_32(struct kvm_vcpu *vcpu, struct kvm_run *run); kvm_vcpu 32 arch/arm64/include/asm/kvm_coproc.h int kvm_handle_cp14_64(struct kvm_vcpu *vcpu, struct kvm_run *run); kvm_vcpu 33 arch/arm64/include/asm/kvm_coproc.h int kvm_handle_cp15_32(struct kvm_vcpu *vcpu, struct kvm_run *run); kvm_vcpu 34 arch/arm64/include/asm/kvm_coproc.h int kvm_handle_cp15_64(struct kvm_vcpu *vcpu, struct kvm_run *run); kvm_vcpu 35 arch/arm64/include/asm/kvm_coproc.h int kvm_handle_sys_reg(struct kvm_vcpu *vcpu, struct kvm_run *run); kvm_vcpu 41 arch/arm64/include/asm/kvm_coproc.h int kvm_arm_copy_sys_reg_indices(struct kvm_vcpu *vcpu, u64 __user *uindices); kvm_vcpu 42 arch/arm64/include/asm/kvm_coproc.h int kvm_arm_sys_reg_get_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *); kvm_vcpu 43 arch/arm64/include/asm/kvm_coproc.h int kvm_arm_sys_reg_set_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *); kvm_vcpu 44 arch/arm64/include/asm/kvm_coproc.h unsigned long kvm_arm_num_sys_reg_descs(struct kvm_vcpu *vcpu); kvm_vcpu 25 arch/arm64/include/asm/kvm_emulate.h unsigned long *vcpu_reg32(const struct kvm_vcpu *vcpu, u8 reg_num); kvm_vcpu 26 arch/arm64/include/asm/kvm_emulate.h unsigned long vcpu_read_spsr32(const struct kvm_vcpu *vcpu); kvm_vcpu 27 arch/arm64/include/asm/kvm_emulate.h void vcpu_write_spsr32(struct kvm_vcpu *vcpu, unsigned long v); kvm_vcpu 29 arch/arm64/include/asm/kvm_emulate.h bool kvm_condition_valid32(const struct kvm_vcpu *vcpu); kvm_vcpu 30 arch/arm64/include/asm/kvm_emulate.h void kvm_skip_instr32(struct kvm_vcpu *vcpu, bool is_wide_instr); kvm_vcpu 32 arch/arm64/include/asm/kvm_emulate.h void kvm_inject_undefined(struct kvm_vcpu *vcpu); kvm_vcpu 33 arch/arm64/include/asm/kvm_emulate.h void kvm_inject_vabt(struct kvm_vcpu *vcpu); kvm_vcpu 34 arch/arm64/include/asm/kvm_emulate.h void kvm_inject_dabt(struct kvm_vcpu *vcpu, unsigned long addr); kvm_vcpu 35 arch/arm64/include/asm/kvm_emulate.h void kvm_inject_pabt(struct kvm_vcpu *vcpu, unsigned long addr); kvm_vcpu 36 arch/arm64/include/asm/kvm_emulate.h void kvm_inject_undef32(struct kvm_vcpu *vcpu); kvm_vcpu 37 arch/arm64/include/asm/kvm_emulate.h void kvm_inject_dabt32(struct kvm_vcpu *vcpu, unsigned long addr); kvm_vcpu 38 arch/arm64/include/asm/kvm_emulate.h void kvm_inject_pabt32(struct kvm_vcpu *vcpu, unsigned long addr); kvm_vcpu 40 arch/arm64/include/asm/kvm_emulate.h static inline bool vcpu_el1_is_32bit(struct kvm_vcpu *vcpu) kvm_vcpu 45 arch/arm64/include/asm/kvm_emulate.h static inline void vcpu_reset_hcr(struct kvm_vcpu *vcpu) kvm_vcpu 75 arch/arm64/include/asm/kvm_emulate.h static inline unsigned long *vcpu_hcr(struct kvm_vcpu *vcpu) kvm_vcpu 80 arch/arm64/include/asm/kvm_emulate.h static inline void vcpu_clear_wfe_traps(struct kvm_vcpu *vcpu) kvm_vcpu 85 arch/arm64/include/asm/kvm_emulate.h static inline void vcpu_set_wfe_traps(struct kvm_vcpu *vcpu) kvm_vcpu 90 arch/arm64/include/asm/kvm_emulate.h static inline void vcpu_ptrauth_enable(struct kvm_vcpu *vcpu) kvm_vcpu 95 arch/arm64/include/asm/kvm_emulate.h static inline void vcpu_ptrauth_disable(struct kvm_vcpu *vcpu) kvm_vcpu 100 arch/arm64/include/asm/kvm_emulate.h static inline unsigned long vcpu_get_vsesr(struct kvm_vcpu *vcpu) kvm_vcpu 105 arch/arm64/include/asm/kvm_emulate.h static inline void vcpu_set_vsesr(struct kvm_vcpu *vcpu, u64 vsesr) kvm_vcpu 110 arch/arm64/include/asm/kvm_emulate.h static inline unsigned long *vcpu_pc(const struct kvm_vcpu *vcpu) kvm_vcpu 115 arch/arm64/include/asm/kvm_emulate.h static inline unsigned long *__vcpu_elr_el1(const struct kvm_vcpu *vcpu) kvm_vcpu 120 arch/arm64/include/asm/kvm_emulate.h static inline unsigned long vcpu_read_elr_el1(const struct kvm_vcpu *vcpu) kvm_vcpu 128 arch/arm64/include/asm/kvm_emulate.h static inline void vcpu_write_elr_el1(const struct kvm_vcpu *vcpu, unsigned long v) kvm_vcpu 136 arch/arm64/include/asm/kvm_emulate.h static inline unsigned long *vcpu_cpsr(const struct kvm_vcpu *vcpu) kvm_vcpu 141 arch/arm64/include/asm/kvm_emulate.h static inline bool vcpu_mode_is_32bit(const struct kvm_vcpu *vcpu) kvm_vcpu 146 arch/arm64/include/asm/kvm_emulate.h static inline bool kvm_condition_valid(const struct kvm_vcpu *vcpu) kvm_vcpu 154 arch/arm64/include/asm/kvm_emulate.h static inline void vcpu_set_thumb(struct kvm_vcpu *vcpu) kvm_vcpu 164 arch/arm64/include/asm/kvm_emulate.h static inline unsigned long vcpu_get_reg(const struct kvm_vcpu *vcpu, kvm_vcpu 170 arch/arm64/include/asm/kvm_emulate.h static inline void vcpu_set_reg(struct kvm_vcpu *vcpu, u8 reg_num, kvm_vcpu 177 arch/arm64/include/asm/kvm_emulate.h static inline unsigned long vcpu_read_spsr(const struct kvm_vcpu *vcpu) kvm_vcpu 188 arch/arm64/include/asm/kvm_emulate.h static inline void vcpu_write_spsr(struct kvm_vcpu *vcpu, unsigned long v) kvm_vcpu 233 arch/arm64/include/asm/kvm_emulate.h static inline bool vcpu_mode_priv(const struct kvm_vcpu *vcpu) kvm_vcpu 247 arch/arm64/include/asm/kvm_emulate.h static inline u32 kvm_vcpu_get_hsr(const struct kvm_vcpu *vcpu) kvm_vcpu 252 arch/arm64/include/asm/kvm_emulate.h static inline int kvm_vcpu_get_condition(const struct kvm_vcpu *vcpu) kvm_vcpu 262 arch/arm64/include/asm/kvm_emulate.h static inline unsigned long kvm_vcpu_get_hfar(const struct kvm_vcpu *vcpu) kvm_vcpu 267 arch/arm64/include/asm/kvm_emulate.h static inline phys_addr_t kvm_vcpu_get_fault_ipa(const struct kvm_vcpu *vcpu) kvm_vcpu 272 arch/arm64/include/asm/kvm_emulate.h static inline u64 kvm_vcpu_get_disr(const struct kvm_vcpu *vcpu) kvm_vcpu 277 arch/arm64/include/asm/kvm_emulate.h static inline u32 kvm_vcpu_hvc_get_imm(const struct kvm_vcpu *vcpu) kvm_vcpu 282 arch/arm64/include/asm/kvm_emulate.h static inline bool kvm_vcpu_dabt_isvalid(const struct kvm_vcpu *vcpu) kvm_vcpu 287 arch/arm64/include/asm/kvm_emulate.h static inline bool kvm_vcpu_dabt_issext(const struct kvm_vcpu *vcpu) kvm_vcpu 292 arch/arm64/include/asm/kvm_emulate.h static inline bool kvm_vcpu_dabt_issf(const struct kvm_vcpu *vcpu) kvm_vcpu 297 arch/arm64/include/asm/kvm_emulate.h static inline int kvm_vcpu_dabt_get_rd(const struct kvm_vcpu *vcpu) kvm_vcpu 302 arch/arm64/include/asm/kvm_emulate.h static inline bool kvm_vcpu_dabt_iss1tw(const struct kvm_vcpu *vcpu) kvm_vcpu 307 arch/arm64/include/asm/kvm_emulate.h static inline bool kvm_vcpu_dabt_iswrite(const struct kvm_vcpu *vcpu) kvm_vcpu 313 arch/arm64/include/asm/kvm_emulate.h static inline bool kvm_vcpu_dabt_is_cm(const struct kvm_vcpu *vcpu) kvm_vcpu 318 arch/arm64/include/asm/kvm_emulate.h static inline int kvm_vcpu_dabt_get_as(const struct kvm_vcpu *vcpu) kvm_vcpu 324 arch/arm64/include/asm/kvm_emulate.h static inline bool kvm_vcpu_trap_il_is32bit(const struct kvm_vcpu *vcpu) kvm_vcpu 329 arch/arm64/include/asm/kvm_emulate.h static inline u8 kvm_vcpu_trap_get_class(const struct kvm_vcpu *vcpu) kvm_vcpu 334 arch/arm64/include/asm/kvm_emulate.h static inline bool kvm_vcpu_trap_is_iabt(const struct kvm_vcpu *vcpu) kvm_vcpu 339 arch/arm64/include/asm/kvm_emulate.h static inline u8 kvm_vcpu_trap_get_fault(const struct kvm_vcpu *vcpu) kvm_vcpu 344 arch/arm64/include/asm/kvm_emulate.h static inline u8 kvm_vcpu_trap_get_fault_type(const struct kvm_vcpu *vcpu) kvm_vcpu 349 arch/arm64/include/asm/kvm_emulate.h static inline bool kvm_vcpu_dabt_isextabt(const struct kvm_vcpu *vcpu) kvm_vcpu 368 arch/arm64/include/asm/kvm_emulate.h static inline int kvm_vcpu_sys_get_rt(struct kvm_vcpu *vcpu) kvm_vcpu 374 arch/arm64/include/asm/kvm_emulate.h static inline bool kvm_is_write_fault(struct kvm_vcpu *vcpu) kvm_vcpu 382 arch/arm64/include/asm/kvm_emulate.h static inline unsigned long kvm_vcpu_get_mpidr_aff(struct kvm_vcpu *vcpu) kvm_vcpu 387 arch/arm64/include/asm/kvm_emulate.h static inline bool kvm_arm_get_vcpu_workaround_2_flag(struct kvm_vcpu *vcpu) kvm_vcpu 392 arch/arm64/include/asm/kvm_emulate.h static inline void kvm_arm_set_vcpu_workaround_2_flag(struct kvm_vcpu *vcpu, kvm_vcpu 401 arch/arm64/include/asm/kvm_emulate.h static inline void kvm_vcpu_set_be(struct kvm_vcpu *vcpu) kvm_vcpu 412 arch/arm64/include/asm/kvm_emulate.h static inline bool kvm_vcpu_is_be(struct kvm_vcpu *vcpu) kvm_vcpu 420 arch/arm64/include/asm/kvm_emulate.h static inline unsigned long vcpu_data_guest_to_host(struct kvm_vcpu *vcpu, kvm_vcpu 451 arch/arm64/include/asm/kvm_emulate.h static inline unsigned long vcpu_data_host_to_guest(struct kvm_vcpu *vcpu, kvm_vcpu 482 arch/arm64/include/asm/kvm_emulate.h static inline void kvm_skip_instr(struct kvm_vcpu *vcpu, bool is_wide_instr) kvm_vcpu 497 arch/arm64/include/asm/kvm_emulate.h static inline void __hyp_text __kvm_skip_instr(struct kvm_vcpu *vcpu) kvm_vcpu 54 arch/arm64/include/asm/kvm_host.h int kvm_reset_vcpu(struct kvm_vcpu *vcpu); kvm_vcpu 55 arch/arm64/include/asm/kvm_host.h void kvm_arch_vcpu_uninit(struct kvm_vcpu *vcpu); kvm_vcpu 222 arch/arm64/include/asm/kvm_host.h struct kvm_vcpu *__hyp_running_vcpu; kvm_vcpu 388 arch/arm64/include/asm/kvm_host.h u64 vcpu_read_sys_reg(const struct kvm_vcpu *vcpu, int reg); kvm_vcpu 389 arch/arm64/include/asm/kvm_host.h void vcpu_write_sys_reg(struct kvm_vcpu *vcpu, u64 val, int reg); kvm_vcpu 418 arch/arm64/include/asm/kvm_host.h unsigned long kvm_arm_num_regs(struct kvm_vcpu *vcpu); kvm_vcpu 419 arch/arm64/include/asm/kvm_host.h int kvm_arm_copy_reg_indices(struct kvm_vcpu *vcpu, u64 __user *indices); kvm_vcpu 420 arch/arm64/include/asm/kvm_host.h int kvm_arm_get_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg); kvm_vcpu 421 arch/arm64/include/asm/kvm_host.h int kvm_arm_set_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg); kvm_vcpu 422 arch/arm64/include/asm/kvm_host.h int __kvm_arm_vcpu_get_events(struct kvm_vcpu *vcpu, kvm_vcpu 425 arch/arm64/include/asm/kvm_host.h int __kvm_arm_vcpu_set_events(struct kvm_vcpu *vcpu, kvm_vcpu 435 arch/arm64/include/asm/kvm_host.h struct kvm_vcpu *kvm_arm_get_running_vcpu(void); kvm_vcpu 436 arch/arm64/include/asm/kvm_host.h struct kvm_vcpu * __percpu *kvm_get_running_vcpus(void); kvm_vcpu 475 arch/arm64/include/asm/kvm_host.h int handle_exit(struct kvm_vcpu *vcpu, struct kvm_run *run, kvm_vcpu 477 arch/arm64/include/asm/kvm_host.h void handle_exit_early(struct kvm_vcpu *vcpu, struct kvm_run *run, kvm_vcpu 483 arch/arm64/include/asm/kvm_host.h void kvm_set_sei_esr(struct kvm_vcpu *vcpu, u64 syndrome); kvm_vcpu 485 arch/arm64/include/asm/kvm_host.h struct kvm_vcpu *kvm_mpidr_to_vcpu(struct kvm *kvm, unsigned long mpidr); kvm_vcpu 545 arch/arm64/include/asm/kvm_host.h void kvm_arm_vcpu_ptrauth_trap(struct kvm_vcpu *vcpu); kvm_vcpu 549 arch/arm64/include/asm/kvm_host.h static inline void kvm_arch_sched_in(struct kvm_vcpu *vcpu, int cpu) {} kvm_vcpu 550 arch/arm64/include/asm/kvm_host.h static inline void kvm_arch_vcpu_block_finish(struct kvm_vcpu *vcpu) {} kvm_vcpu 553 arch/arm64/include/asm/kvm_host.h void kvm_arm_setup_debug(struct kvm_vcpu *vcpu); kvm_vcpu 554 arch/arm64/include/asm/kvm_host.h void kvm_arm_clear_debug(struct kvm_vcpu *vcpu); kvm_vcpu 555 arch/arm64/include/asm/kvm_host.h void kvm_arm_reset_debug_ptr(struct kvm_vcpu *vcpu); kvm_vcpu 556 arch/arm64/include/asm/kvm_host.h int kvm_arm_vcpu_arch_set_attr(struct kvm_vcpu *vcpu, kvm_vcpu 558 arch/arm64/include/asm/kvm_host.h int kvm_arm_vcpu_arch_get_attr(struct kvm_vcpu *vcpu, kvm_vcpu 560 arch/arm64/include/asm/kvm_host.h int kvm_arm_vcpu_arch_has_attr(struct kvm_vcpu *vcpu, kvm_vcpu 566 arch/arm64/include/asm/kvm_host.h int kvm_arch_vcpu_run_map_fp(struct kvm_vcpu *vcpu); kvm_vcpu 567 arch/arm64/include/asm/kvm_host.h void kvm_arch_vcpu_load_fp(struct kvm_vcpu *vcpu); kvm_vcpu 568 arch/arm64/include/asm/kvm_host.h void kvm_arch_vcpu_ctxsync_fp(struct kvm_vcpu *vcpu); kvm_vcpu 569 arch/arm64/include/asm/kvm_host.h void kvm_arch_vcpu_put_fp(struct kvm_vcpu *vcpu); kvm_vcpu 577 arch/arm64/include/asm/kvm_host.h static inline int kvm_arch_vcpu_run_pid_change(struct kvm_vcpu *vcpu) kvm_vcpu 585 arch/arm64/include/asm/kvm_host.h void kvm_vcpu_pmu_restore_guest(struct kvm_vcpu *vcpu); kvm_vcpu 586 arch/arm64/include/asm/kvm_host.h void kvm_vcpu_pmu_restore_host(struct kvm_vcpu *vcpu); kvm_vcpu 665 arch/arm64/include/asm/kvm_host.h void kvm_vcpu_load_sysregs(struct kvm_vcpu *vcpu); kvm_vcpu 666 arch/arm64/include/asm/kvm_host.h void kvm_vcpu_put_sysregs(struct kvm_vcpu *vcpu); kvm_vcpu 676 arch/arm64/include/asm/kvm_host.h int kvm_arm_vcpu_finalize(struct kvm_vcpu *vcpu, int feature); kvm_vcpu 677 arch/arm64/include/asm/kvm_host.h bool kvm_arm_vcpu_is_finalized(struct kvm_vcpu *vcpu); kvm_vcpu 50 arch/arm64/include/asm/kvm_hyp.h int __vgic_v2_perform_cpuif_access(struct kvm_vcpu *vcpu); kvm_vcpu 52 arch/arm64/include/asm/kvm_hyp.h void __vgic_v3_save_state(struct kvm_vcpu *vcpu); kvm_vcpu 53 arch/arm64/include/asm/kvm_hyp.h void __vgic_v3_restore_state(struct kvm_vcpu *vcpu); kvm_vcpu 54 arch/arm64/include/asm/kvm_hyp.h void __vgic_v3_activate_traps(struct kvm_vcpu *vcpu); kvm_vcpu 55 arch/arm64/include/asm/kvm_hyp.h void __vgic_v3_deactivate_traps(struct kvm_vcpu *vcpu); kvm_vcpu 56 arch/arm64/include/asm/kvm_hyp.h void __vgic_v3_save_aprs(struct kvm_vcpu *vcpu); kvm_vcpu 57 arch/arm64/include/asm/kvm_hyp.h void __vgic_v3_restore_aprs(struct kvm_vcpu *vcpu); kvm_vcpu 58 arch/arm64/include/asm/kvm_hyp.h int __vgic_v3_perform_cpuif_access(struct kvm_vcpu *vcpu); kvm_vcpu 60 arch/arm64/include/asm/kvm_hyp.h void __timer_enable_traps(struct kvm_vcpu *vcpu); kvm_vcpu 61 arch/arm64/include/asm/kvm_hyp.h void __timer_disable_traps(struct kvm_vcpu *vcpu); kvm_vcpu 69 arch/arm64/include/asm/kvm_hyp.h void __sysreg32_save_state(struct kvm_vcpu *vcpu); kvm_vcpu 70 arch/arm64/include/asm/kvm_hyp.h void __sysreg32_restore_state(struct kvm_vcpu *vcpu); kvm_vcpu 72 arch/arm64/include/asm/kvm_hyp.h void __debug_switch_to_guest(struct kvm_vcpu *vcpu); kvm_vcpu 73 arch/arm64/include/asm/kvm_hyp.h void __debug_switch_to_host(struct kvm_vcpu *vcpu); kvm_vcpu 78 arch/arm64/include/asm/kvm_hyp.h void activate_traps_vhe_load(struct kvm_vcpu *vcpu); kvm_vcpu 81 arch/arm64/include/asm/kvm_hyp.h u64 __guest_enter(struct kvm_vcpu *vcpu, struct kvm_cpu_context *host_ctxt); kvm_vcpu 23 arch/arm64/include/asm/kvm_mmio.h int kvm_handle_mmio_return(struct kvm_vcpu *vcpu, struct kvm_run *run); kvm_vcpu 24 arch/arm64/include/asm/kvm_mmio.h int io_mem_abort(struct kvm_vcpu *vcpu, struct kvm_run *run, kvm_vcpu 161 arch/arm64/include/asm/kvm_mmu.h int kvm_handle_guest_abort(struct kvm_vcpu *vcpu, struct kvm_run *run); kvm_vcpu 163 arch/arm64/include/asm/kvm_mmu.h void kvm_mmu_free_memory_caches(struct kvm_vcpu *vcpu); kvm_vcpu 305 arch/arm64/include/asm/kvm_mmu.h static inline bool vcpu_has_cache_enabled(struct kvm_vcpu *vcpu) kvm_vcpu 367 arch/arm64/include/asm/kvm_mmu.h void kvm_set_way_flush(struct kvm_vcpu *vcpu); kvm_vcpu 368 arch/arm64/include/asm/kvm_mmu.h void kvm_toggle_cache(struct kvm_vcpu *vcpu, bool was_enabled); kvm_vcpu 92 arch/arm64/kernel/asm-offsets.c DEFINE(VCPU_CONTEXT, offsetof(struct kvm_vcpu, arch.ctxt)); kvm_vcpu 93 arch/arm64/kernel/asm-offsets.c DEFINE(VCPU_FAULT_DISR, offsetof(struct kvm_vcpu, arch.fault.disr_el1)); kvm_vcpu 94 arch/arm64/kernel/asm-offsets.c DEFINE(VCPU_WORKAROUND_FLAGS, offsetof(struct kvm_vcpu, arch.workaround_flags)); kvm_vcpu 95 arch/arm64/kernel/asm-offsets.c DEFINE(VCPU_HCR_EL2, offsetof(struct kvm_vcpu, arch.hcr_el2)); kvm_vcpu 36 arch/arm64/kvm/debug.c static void save_guest_debug_regs(struct kvm_vcpu *vcpu) kvm_vcpu 46 arch/arm64/kvm/debug.c static void restore_guest_debug_regs(struct kvm_vcpu *vcpu) kvm_vcpu 75 arch/arm64/kvm/debug.c void kvm_arm_reset_debug_ptr(struct kvm_vcpu *vcpu) kvm_vcpu 101 arch/arm64/kvm/debug.c void kvm_arm_setup_debug(struct kvm_vcpu *vcpu) kvm_vcpu 208 arch/arm64/kvm/debug.c void kvm_arm_clear_debug(struct kvm_vcpu *vcpu) kvm_vcpu 27 arch/arm64/kvm/fpsimd.c int kvm_arch_vcpu_run_map_fp(struct kvm_vcpu *vcpu) kvm_vcpu 62 arch/arm64/kvm/fpsimd.c void kvm_arch_vcpu_load_fp(struct kvm_vcpu *vcpu) kvm_vcpu 84 arch/arm64/kvm/fpsimd.c void kvm_arch_vcpu_ctxsync_fp(struct kvm_vcpu *vcpu) kvm_vcpu 104 arch/arm64/kvm/fpsimd.c void kvm_arch_vcpu_put_fp(struct kvm_vcpu *vcpu) kvm_vcpu 34 arch/arm64/kvm/guest.c #define VCPU_STAT(x) { #x, offsetof(struct kvm_vcpu, stat.x), KVM_STAT_VCPU } kvm_vcpu 46 arch/arm64/kvm/guest.c int kvm_arch_vcpu_setup(struct kvm_vcpu *vcpu) kvm_vcpu 62 arch/arm64/kvm/guest.c static int core_reg_size_from_offset(const struct kvm_vcpu *vcpu, u64 off) kvm_vcpu 107 arch/arm64/kvm/guest.c static int validate_core_offset(const struct kvm_vcpu *vcpu, kvm_vcpu 122 arch/arm64/kvm/guest.c static int get_core_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg) kvm_vcpu 150 arch/arm64/kvm/guest.c static int set_core_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg) kvm_vcpu 220 arch/arm64/kvm/guest.c static int get_sve_vls(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg) kvm_vcpu 244 arch/arm64/kvm/guest.c static int set_sve_vls(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg) kvm_vcpu 326 arch/arm64/kvm/guest.c struct kvm_vcpu *vcpu, kvm_vcpu 392 arch/arm64/kvm/guest.c static int get_sve_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg) kvm_vcpu 418 arch/arm64/kvm/guest.c static int set_sve_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg) kvm_vcpu 443 arch/arm64/kvm/guest.c int kvm_arch_vcpu_ioctl_get_regs(struct kvm_vcpu *vcpu, struct kvm_regs *regs) kvm_vcpu 448 arch/arm64/kvm/guest.c int kvm_arch_vcpu_ioctl_set_regs(struct kvm_vcpu *vcpu, struct kvm_regs *regs) kvm_vcpu 453 arch/arm64/kvm/guest.c static int copy_core_reg_indices(const struct kvm_vcpu *vcpu, kvm_vcpu 496 arch/arm64/kvm/guest.c static unsigned long num_core_regs(const struct kvm_vcpu *vcpu) kvm_vcpu 518 arch/arm64/kvm/guest.c static int copy_timer_indices(struct kvm_vcpu *vcpu, u64 __user *uindices) kvm_vcpu 532 arch/arm64/kvm/guest.c static int set_timer_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg) kvm_vcpu 545 arch/arm64/kvm/guest.c static int get_timer_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg) kvm_vcpu 554 arch/arm64/kvm/guest.c static unsigned long num_sve_regs(const struct kvm_vcpu *vcpu) kvm_vcpu 568 arch/arm64/kvm/guest.c static int copy_sve_reg_indices(const struct kvm_vcpu *vcpu, kvm_vcpu 620 arch/arm64/kvm/guest.c unsigned long kvm_arm_num_regs(struct kvm_vcpu *vcpu) kvm_vcpu 638 arch/arm64/kvm/guest.c int kvm_arm_copy_reg_indices(struct kvm_vcpu *vcpu, u64 __user *uindices) kvm_vcpu 665 arch/arm64/kvm/guest.c int kvm_arm_get_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg) kvm_vcpu 683 arch/arm64/kvm/guest.c int kvm_arm_set_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg) kvm_vcpu 701 arch/arm64/kvm/guest.c int kvm_arch_vcpu_ioctl_get_sregs(struct kvm_vcpu *vcpu, kvm_vcpu 707 arch/arm64/kvm/guest.c int kvm_arch_vcpu_ioctl_set_sregs(struct kvm_vcpu *vcpu, kvm_vcpu 713 arch/arm64/kvm/guest.c int __kvm_arm_vcpu_get_events(struct kvm_vcpu *vcpu, kvm_vcpu 725 arch/arm64/kvm/guest.c int __kvm_arm_vcpu_set_events(struct kvm_vcpu *vcpu, kvm_vcpu 796 arch/arm64/kvm/guest.c int kvm_arch_vcpu_ioctl_get_fpu(struct kvm_vcpu *vcpu, struct kvm_fpu *fpu) kvm_vcpu 801 arch/arm64/kvm/guest.c int kvm_arch_vcpu_ioctl_set_fpu(struct kvm_vcpu *vcpu, struct kvm_fpu *fpu) kvm_vcpu 806 arch/arm64/kvm/guest.c int kvm_arch_vcpu_ioctl_translate(struct kvm_vcpu *vcpu, kvm_vcpu 827 arch/arm64/kvm/guest.c int kvm_arch_vcpu_ioctl_set_guest_debug(struct kvm_vcpu *vcpu, kvm_vcpu 856 arch/arm64/kvm/guest.c int kvm_arm_vcpu_arch_set_attr(struct kvm_vcpu *vcpu, kvm_vcpu 876 arch/arm64/kvm/guest.c int kvm_arm_vcpu_arch_get_attr(struct kvm_vcpu *vcpu, kvm_vcpu 896 arch/arm64/kvm/guest.c int kvm_arm_vcpu_arch_has_attr(struct kvm_vcpu *vcpu, kvm_vcpu 28 arch/arm64/kvm/handle_exit.c typedef int (*exit_handle_fn)(struct kvm_vcpu *, struct kvm_run *); kvm_vcpu 30 arch/arm64/kvm/handle_exit.c static void kvm_handle_guest_serror(struct kvm_vcpu *vcpu, u32 esr) kvm_vcpu 36 arch/arm64/kvm/handle_exit.c static int handle_hvc(struct kvm_vcpu *vcpu, struct kvm_run *run) kvm_vcpu 53 arch/arm64/kvm/handle_exit.c static int handle_smc(struct kvm_vcpu *vcpu, struct kvm_run *run) kvm_vcpu 72 arch/arm64/kvm/handle_exit.c static int handle_no_fpsimd(struct kvm_vcpu *vcpu, struct kvm_run *run) kvm_vcpu 90 arch/arm64/kvm/handle_exit.c static int kvm_handle_wfx(struct kvm_vcpu *vcpu, struct kvm_run *run) kvm_vcpu 120 arch/arm64/kvm/handle_exit.c static int kvm_handle_guest_debug(struct kvm_vcpu *vcpu, struct kvm_run *run) kvm_vcpu 147 arch/arm64/kvm/handle_exit.c static int kvm_handle_unknown_ec(struct kvm_vcpu *vcpu, struct kvm_run *run) kvm_vcpu 158 arch/arm64/kvm/handle_exit.c static int handle_sve(struct kvm_vcpu *vcpu, struct kvm_run *run) kvm_vcpu 169 arch/arm64/kvm/handle_exit.c void kvm_arm_vcpu_ptrauth_trap(struct kvm_vcpu *vcpu) kvm_vcpu 181 arch/arm64/kvm/handle_exit.c static int kvm_handle_ptrauth(struct kvm_vcpu *vcpu, struct kvm_run *run) kvm_vcpu 212 arch/arm64/kvm/handle_exit.c static exit_handle_fn kvm_get_exit_handler(struct kvm_vcpu *vcpu) kvm_vcpu 226 arch/arm64/kvm/handle_exit.c static int handle_trap_exceptions(struct kvm_vcpu *vcpu, struct kvm_run *run) kvm_vcpu 251 arch/arm64/kvm/handle_exit.c int handle_exit(struct kvm_vcpu *vcpu, struct kvm_run *run, kvm_vcpu 303 arch/arm64/kvm/handle_exit.c void handle_exit_early(struct kvm_vcpu *vcpu, struct kvm_run *run, kvm_vcpu 132 arch/arm64/kvm/hyp/debug-sr.c static void __hyp_text __debug_save_state(struct kvm_vcpu *vcpu, kvm_vcpu 151 arch/arm64/kvm/hyp/debug-sr.c static void __hyp_text __debug_restore_state(struct kvm_vcpu *vcpu, kvm_vcpu 171 arch/arm64/kvm/hyp/debug-sr.c void __hyp_text __debug_switch_to_guest(struct kvm_vcpu *vcpu) kvm_vcpu 197 arch/arm64/kvm/hyp/debug-sr.c void __hyp_text __debug_switch_to_host(struct kvm_vcpu *vcpu) kvm_vcpu 29 arch/arm64/kvm/hyp/switch.c static bool __hyp_text update_fp_enabled(struct kvm_vcpu *vcpu) kvm_vcpu 47 arch/arm64/kvm/hyp/switch.c static void __hyp_text __fpsimd_save_fpexc32(struct kvm_vcpu *vcpu) kvm_vcpu 55 arch/arm64/kvm/hyp/switch.c static void __hyp_text __activate_traps_fpsimd32(struct kvm_vcpu *vcpu) kvm_vcpu 72 arch/arm64/kvm/hyp/switch.c static void __hyp_text __activate_traps_common(struct kvm_vcpu *vcpu) kvm_vcpu 94 arch/arm64/kvm/hyp/switch.c static void activate_traps_vhe(struct kvm_vcpu *vcpu) kvm_vcpu 115 arch/arm64/kvm/hyp/switch.c static void __hyp_text __activate_traps_nvhe(struct kvm_vcpu *vcpu) kvm_vcpu 131 arch/arm64/kvm/hyp/switch.c static void __hyp_text __activate_traps(struct kvm_vcpu *vcpu) kvm_vcpu 180 arch/arm64/kvm/hyp/switch.c static void __hyp_text __deactivate_traps(struct kvm_vcpu *vcpu) kvm_vcpu 199 arch/arm64/kvm/hyp/switch.c void activate_traps_vhe_load(struct kvm_vcpu *vcpu) kvm_vcpu 222 arch/arm64/kvm/hyp/switch.c static void __hyp_text __deactivate_vm(struct kvm_vcpu *vcpu) kvm_vcpu 228 arch/arm64/kvm/hyp/switch.c static void __hyp_text __hyp_vgic_save_state(struct kvm_vcpu *vcpu) kvm_vcpu 237 arch/arm64/kvm/hyp/switch.c static void __hyp_text __hyp_vgic_restore_state(struct kvm_vcpu *vcpu) kvm_vcpu 274 arch/arm64/kvm/hyp/switch.c static bool __hyp_text __populate_fault_info(struct kvm_vcpu *vcpu) kvm_vcpu 314 arch/arm64/kvm/hyp/switch.c static bool __hyp_text __hyp_handle_fpsimd(struct kvm_vcpu *vcpu) kvm_vcpu 396 arch/arm64/kvm/hyp/switch.c static bool __hyp_text handle_tx2_tvm(struct kvm_vcpu *vcpu) kvm_vcpu 456 arch/arm64/kvm/hyp/switch.c static bool __hyp_text fixup_guest_exit(struct kvm_vcpu *vcpu, u64 *exit_code) kvm_vcpu 525 arch/arm64/kvm/hyp/switch.c static inline bool __hyp_text __needs_ssbd_off(struct kvm_vcpu *vcpu) kvm_vcpu 533 arch/arm64/kvm/hyp/switch.c static void __hyp_text __set_guest_arch_workaround_state(struct kvm_vcpu *vcpu) kvm_vcpu 546 arch/arm64/kvm/hyp/switch.c static void __hyp_text __set_host_arch_workaround_state(struct kvm_vcpu *vcpu) kvm_vcpu 597 arch/arm64/kvm/hyp/switch.c int kvm_vcpu_run_vhe(struct kvm_vcpu *vcpu) kvm_vcpu 653 arch/arm64/kvm/hyp/switch.c int __hyp_text __kvm_vcpu_run_nvhe(struct kvm_vcpu *vcpu) kvm_vcpu 740 arch/arm64/kvm/hyp/switch.c struct kvm_vcpu *vcpu; kvm_vcpu 768 arch/arm64/kvm/hyp/switch.c struct kvm_vcpu *vcpu; kvm_vcpu 191 arch/arm64/kvm/hyp/sysreg-sr.c void __hyp_text __sysreg32_save_state(struct kvm_vcpu *vcpu) kvm_vcpu 213 arch/arm64/kvm/hyp/sysreg-sr.c void __hyp_text __sysreg32_restore_state(struct kvm_vcpu *vcpu) kvm_vcpu 246 arch/arm64/kvm/hyp/sysreg-sr.c void kvm_vcpu_load_sysregs(struct kvm_vcpu *vcpu) kvm_vcpu 282 arch/arm64/kvm/hyp/sysreg-sr.c void kvm_vcpu_put_sysregs(struct kvm_vcpu *vcpu) kvm_vcpu 185 arch/arm64/kvm/hyp/tlb.c void __hyp_text __kvm_tlb_flush_local_vmid(struct kvm_vcpu *vcpu) kvm_vcpu 16 arch/arm64/kvm/hyp/vgic-v2-cpuif-proxy.c static bool __hyp_text __is_be(struct kvm_vcpu *vcpu) kvm_vcpu 35 arch/arm64/kvm/hyp/vgic-v2-cpuif-proxy.c int __hyp_text __vgic_v2_perform_cpuif_access(struct kvm_vcpu *vcpu) kvm_vcpu 29 arch/arm64/kvm/inject_fault.c static u64 get_except_vector(struct kvm_vcpu *vcpu, enum exception_type type) kvm_vcpu 62 arch/arm64/kvm/inject_fault.c static unsigned long get_except64_pstate(struct kvm_vcpu *vcpu) kvm_vcpu 113 arch/arm64/kvm/inject_fault.c static void inject_abt64(struct kvm_vcpu *vcpu, bool is_iabt, unsigned long addr) kvm_vcpu 149 arch/arm64/kvm/inject_fault.c static void inject_undef64(struct kvm_vcpu *vcpu) kvm_vcpu 178 arch/arm64/kvm/inject_fault.c void kvm_inject_dabt(struct kvm_vcpu *vcpu, unsigned long addr) kvm_vcpu 194 arch/arm64/kvm/inject_fault.c void kvm_inject_pabt(struct kvm_vcpu *vcpu, unsigned long addr) kvm_vcpu 208 arch/arm64/kvm/inject_fault.c void kvm_inject_undefined(struct kvm_vcpu *vcpu) kvm_vcpu 216 arch/arm64/kvm/inject_fault.c void kvm_set_sei_esr(struct kvm_vcpu *vcpu, u64 esr) kvm_vcpu 234 arch/arm64/kvm/inject_fault.c void kvm_inject_vabt(struct kvm_vcpu *vcpu) kvm_vcpu 164 arch/arm64/kvm/pmu.c void kvm_vcpu_pmu_restore_guest(struct kvm_vcpu *vcpu) kvm_vcpu 185 arch/arm64/kvm/pmu.c void kvm_vcpu_pmu_restore_host(struct kvm_vcpu *vcpu) kvm_vcpu 101 arch/arm64/kvm/regmap.c unsigned long *vcpu_reg32(const struct kvm_vcpu *vcpu, u8 reg_num) kvm_vcpu 133 arch/arm64/kvm/regmap.c static int vcpu_spsr32_mode(const struct kvm_vcpu *vcpu) kvm_vcpu 146 arch/arm64/kvm/regmap.c unsigned long vcpu_read_spsr32(const struct kvm_vcpu *vcpu) kvm_vcpu 169 arch/arm64/kvm/regmap.c void vcpu_write_spsr32(struct kvm_vcpu *vcpu, unsigned long v) kvm_vcpu 133 arch/arm64/kvm/reset.c static int kvm_vcpu_enable_sve(struct kvm_vcpu *vcpu) kvm_vcpu 158 arch/arm64/kvm/reset.c static int kvm_vcpu_finalize_sve(struct kvm_vcpu *vcpu) kvm_vcpu 183 arch/arm64/kvm/reset.c int kvm_arm_vcpu_finalize(struct kvm_vcpu *vcpu, int feature) kvm_vcpu 199 arch/arm64/kvm/reset.c bool kvm_arm_vcpu_is_finalized(struct kvm_vcpu *vcpu) kvm_vcpu 207 arch/arm64/kvm/reset.c void kvm_arch_vcpu_uninit(struct kvm_vcpu *vcpu) kvm_vcpu 212 arch/arm64/kvm/reset.c static void kvm_vcpu_reset_sve(struct kvm_vcpu *vcpu) kvm_vcpu 218 arch/arm64/kvm/reset.c static int kvm_vcpu_enable_ptrauth(struct kvm_vcpu *vcpu) kvm_vcpu 258 arch/arm64/kvm/reset.c int kvm_reset_vcpu(struct kvm_vcpu *vcpu) kvm_vcpu 48 arch/arm64/kvm/sys_regs.c static bool read_from_write_only(struct kvm_vcpu *vcpu, kvm_vcpu 58 arch/arm64/kvm/sys_regs.c static bool write_to_read_only(struct kvm_vcpu *vcpu, kvm_vcpu 68 arch/arm64/kvm/sys_regs.c u64 vcpu_read_sys_reg(const struct kvm_vcpu *vcpu, int reg) kvm_vcpu 112 arch/arm64/kvm/sys_regs.c void vcpu_write_sys_reg(struct kvm_vcpu *vcpu, u64 val, int reg) kvm_vcpu 179 arch/arm64/kvm/sys_regs.c static bool access_dcsw(struct kvm_vcpu *vcpu, kvm_vcpu 204 arch/arm64/kvm/sys_regs.c static bool access_vm_reg(struct kvm_vcpu *vcpu, kvm_vcpu 240 arch/arm64/kvm/sys_regs.c static bool access_gic_sgi(struct kvm_vcpu *vcpu, kvm_vcpu 285 arch/arm64/kvm/sys_regs.c static bool access_gic_sre(struct kvm_vcpu *vcpu, kvm_vcpu 296 arch/arm64/kvm/sys_regs.c static bool trap_raz_wi(struct kvm_vcpu *vcpu, kvm_vcpu 312 arch/arm64/kvm/sys_regs.c static bool trap_loregion(struct kvm_vcpu *vcpu, kvm_vcpu 331 arch/arm64/kvm/sys_regs.c static bool trap_oslsr_el1(struct kvm_vcpu *vcpu, kvm_vcpu 343 arch/arm64/kvm/sys_regs.c static bool trap_dbgauthstatus_el1(struct kvm_vcpu *vcpu, kvm_vcpu 382 arch/arm64/kvm/sys_regs.c static bool trap_debug_regs(struct kvm_vcpu *vcpu, kvm_vcpu 407 arch/arm64/kvm/sys_regs.c static void reg_to_dbg(struct kvm_vcpu *vcpu, kvm_vcpu 422 arch/arm64/kvm/sys_regs.c static void dbg_to_reg(struct kvm_vcpu *vcpu, kvm_vcpu 431 arch/arm64/kvm/sys_regs.c static bool trap_bvr(struct kvm_vcpu *vcpu, kvm_vcpu 447 arch/arm64/kvm/sys_regs.c static int set_bvr(struct kvm_vcpu *vcpu, const struct sys_reg_desc *rd, kvm_vcpu 457 arch/arm64/kvm/sys_regs.c static int get_bvr(struct kvm_vcpu *vcpu, const struct sys_reg_desc *rd, kvm_vcpu 467 arch/arm64/kvm/sys_regs.c static void reset_bvr(struct kvm_vcpu *vcpu, kvm_vcpu 473 arch/arm64/kvm/sys_regs.c static bool trap_bcr(struct kvm_vcpu *vcpu, kvm_vcpu 489 arch/arm64/kvm/sys_regs.c static int set_bcr(struct kvm_vcpu *vcpu, const struct sys_reg_desc *rd, kvm_vcpu 500 arch/arm64/kvm/sys_regs.c static int get_bcr(struct kvm_vcpu *vcpu, const struct sys_reg_desc *rd, kvm_vcpu 510 arch/arm64/kvm/sys_regs.c static void reset_bcr(struct kvm_vcpu *vcpu, kvm_vcpu 516 arch/arm64/kvm/sys_regs.c static bool trap_wvr(struct kvm_vcpu *vcpu, kvm_vcpu 533 arch/arm64/kvm/sys_regs.c static int set_wvr(struct kvm_vcpu *vcpu, const struct sys_reg_desc *rd, kvm_vcpu 543 arch/arm64/kvm/sys_regs.c static int get_wvr(struct kvm_vcpu *vcpu, const struct sys_reg_desc *rd, kvm_vcpu 553 arch/arm64/kvm/sys_regs.c static void reset_wvr(struct kvm_vcpu *vcpu, kvm_vcpu 559 arch/arm64/kvm/sys_regs.c static bool trap_wcr(struct kvm_vcpu *vcpu, kvm_vcpu 575 arch/arm64/kvm/sys_regs.c static int set_wcr(struct kvm_vcpu *vcpu, const struct sys_reg_desc *rd, kvm_vcpu 585 arch/arm64/kvm/sys_regs.c static int get_wcr(struct kvm_vcpu *vcpu, const struct sys_reg_desc *rd, kvm_vcpu 595 arch/arm64/kvm/sys_regs.c static void reset_wcr(struct kvm_vcpu *vcpu, kvm_vcpu 601 arch/arm64/kvm/sys_regs.c static void reset_amair_el1(struct kvm_vcpu *vcpu, const struct sys_reg_desc *r) kvm_vcpu 607 arch/arm64/kvm/sys_regs.c static void reset_mpidr(struct kvm_vcpu *vcpu, const struct sys_reg_desc *r) kvm_vcpu 624 arch/arm64/kvm/sys_regs.c static void reset_pmcr(struct kvm_vcpu *vcpu, const struct sys_reg_desc *r) kvm_vcpu 640 arch/arm64/kvm/sys_regs.c static bool check_pmu_access_disabled(struct kvm_vcpu *vcpu, u64 flags) kvm_vcpu 651 arch/arm64/kvm/sys_regs.c static bool pmu_access_el0_disabled(struct kvm_vcpu *vcpu) kvm_vcpu 656 arch/arm64/kvm/sys_regs.c static bool pmu_write_swinc_el0_disabled(struct kvm_vcpu *vcpu) kvm_vcpu 661 arch/arm64/kvm/sys_regs.c static bool pmu_access_cycle_counter_el0_disabled(struct kvm_vcpu *vcpu) kvm_vcpu 666 arch/arm64/kvm/sys_regs.c static bool pmu_access_event_counter_el0_disabled(struct kvm_vcpu *vcpu) kvm_vcpu 671 arch/arm64/kvm/sys_regs.c static bool access_pmcr(struct kvm_vcpu *vcpu, struct sys_reg_params *p, kvm_vcpu 702 arch/arm64/kvm/sys_regs.c static bool access_pmselr(struct kvm_vcpu *vcpu, struct sys_reg_params *p, kvm_vcpu 721 arch/arm64/kvm/sys_regs.c static bool access_pmceid(struct kvm_vcpu *vcpu, struct sys_reg_params *p, kvm_vcpu 744 arch/arm64/kvm/sys_regs.c static bool pmu_counter_idx_valid(struct kvm_vcpu *vcpu, u64 idx) kvm_vcpu 758 arch/arm64/kvm/sys_regs.c static bool access_pmu_evcntr(struct kvm_vcpu *vcpu, kvm_vcpu 815 arch/arm64/kvm/sys_regs.c static bool access_pmu_evtyper(struct kvm_vcpu *vcpu, struct sys_reg_params *p, kvm_vcpu 855 arch/arm64/kvm/sys_regs.c static bool access_pmcnten(struct kvm_vcpu *vcpu, struct sys_reg_params *p, kvm_vcpu 886 arch/arm64/kvm/sys_regs.c static bool access_pminten(struct kvm_vcpu *vcpu, struct sys_reg_params *p, kvm_vcpu 915 arch/arm64/kvm/sys_regs.c static bool access_pmovs(struct kvm_vcpu *vcpu, struct sys_reg_params *p, kvm_vcpu 940 arch/arm64/kvm/sys_regs.c static bool access_pmswinc(struct kvm_vcpu *vcpu, struct sys_reg_params *p, kvm_vcpu 959 arch/arm64/kvm/sys_regs.c static bool access_pmuserenr(struct kvm_vcpu *vcpu, struct sys_reg_params *p, kvm_vcpu 1006 arch/arm64/kvm/sys_regs.c static bool trap_ptrauth(struct kvm_vcpu *vcpu, kvm_vcpu 1023 arch/arm64/kvm/sys_regs.c static unsigned int ptrauth_visibility(const struct kvm_vcpu *vcpu, kvm_vcpu 1037 arch/arm64/kvm/sys_regs.c static bool access_arch_timer(struct kvm_vcpu *vcpu, kvm_vcpu 1074 arch/arm64/kvm/sys_regs.c static u64 read_id_reg(const struct kvm_vcpu *vcpu, kvm_vcpu 1095 arch/arm64/kvm/sys_regs.c static bool __access_id_reg(struct kvm_vcpu *vcpu, kvm_vcpu 1107 arch/arm64/kvm/sys_regs.c static bool access_id_reg(struct kvm_vcpu *vcpu, kvm_vcpu 1114 arch/arm64/kvm/sys_regs.c static bool access_raz_id_reg(struct kvm_vcpu *vcpu, kvm_vcpu 1126 arch/arm64/kvm/sys_regs.c static unsigned int sve_visibility(const struct kvm_vcpu *vcpu, kvm_vcpu 1136 arch/arm64/kvm/sys_regs.c static unsigned int sve_id_visibility(const struct kvm_vcpu *vcpu, kvm_vcpu 1146 arch/arm64/kvm/sys_regs.c static u64 guest_id_aa64zfr0_el1(const struct kvm_vcpu *vcpu) kvm_vcpu 1154 arch/arm64/kvm/sys_regs.c static bool access_id_aa64zfr0_el1(struct kvm_vcpu *vcpu, kvm_vcpu 1165 arch/arm64/kvm/sys_regs.c static int get_id_aa64zfr0_el1(struct kvm_vcpu *vcpu, kvm_vcpu 1178 arch/arm64/kvm/sys_regs.c static int set_id_aa64zfr0_el1(struct kvm_vcpu *vcpu, kvm_vcpu 1207 arch/arm64/kvm/sys_regs.c static int __get_id_reg(const struct kvm_vcpu *vcpu, kvm_vcpu 1217 arch/arm64/kvm/sys_regs.c static int __set_id_reg(const struct kvm_vcpu *vcpu, kvm_vcpu 1236 arch/arm64/kvm/sys_regs.c static int get_id_reg(struct kvm_vcpu *vcpu, const struct sys_reg_desc *rd, kvm_vcpu 1242 arch/arm64/kvm/sys_regs.c static int set_id_reg(struct kvm_vcpu *vcpu, const struct sys_reg_desc *rd, kvm_vcpu 1248 arch/arm64/kvm/sys_regs.c static int get_raz_id_reg(struct kvm_vcpu *vcpu, const struct sys_reg_desc *rd, kvm_vcpu 1254 arch/arm64/kvm/sys_regs.c static int set_raz_id_reg(struct kvm_vcpu *vcpu, const struct sys_reg_desc *rd, kvm_vcpu 1260 arch/arm64/kvm/sys_regs.c static bool access_ctr(struct kvm_vcpu *vcpu, struct sys_reg_params *p, kvm_vcpu 1270 arch/arm64/kvm/sys_regs.c static bool access_clidr(struct kvm_vcpu *vcpu, struct sys_reg_params *p, kvm_vcpu 1280 arch/arm64/kvm/sys_regs.c static bool access_csselr(struct kvm_vcpu *vcpu, struct sys_reg_params *p, kvm_vcpu 1296 arch/arm64/kvm/sys_regs.c static bool access_ccsidr(struct kvm_vcpu *vcpu, struct sys_reg_params *p, kvm_vcpu 1653 arch/arm64/kvm/sys_regs.c static bool trap_dbgidr(struct kvm_vcpu *vcpu, kvm_vcpu 1672 arch/arm64/kvm/sys_regs.c static bool trap_debug32(struct kvm_vcpu *vcpu, kvm_vcpu 1697 arch/arm64/kvm/sys_regs.c static bool trap_xvr(struct kvm_vcpu *vcpu, kvm_vcpu 2027 arch/arm64/kvm/sys_regs.c int kvm_handle_cp14_load_store(struct kvm_vcpu *vcpu, struct kvm_run *run) kvm_vcpu 2033 arch/arm64/kvm/sys_regs.c static void perform_access(struct kvm_vcpu *vcpu, kvm_vcpu 2067 arch/arm64/kvm/sys_regs.c static int emulate_cp(struct kvm_vcpu *vcpu, kvm_vcpu 2088 arch/arm64/kvm/sys_regs.c static void unhandled_cp_access(struct kvm_vcpu *vcpu, kvm_vcpu 2118 arch/arm64/kvm/sys_regs.c static int kvm_handle_cp_64(struct kvm_vcpu *vcpu, kvm_vcpu 2175 arch/arm64/kvm/sys_regs.c static int kvm_handle_cp_32(struct kvm_vcpu *vcpu, kvm_vcpu 2206 arch/arm64/kvm/sys_regs.c int kvm_handle_cp15_64(struct kvm_vcpu *vcpu, struct kvm_run *run) kvm_vcpu 2217 arch/arm64/kvm/sys_regs.c int kvm_handle_cp15_32(struct kvm_vcpu *vcpu, struct kvm_run *run) kvm_vcpu 2228 arch/arm64/kvm/sys_regs.c int kvm_handle_cp14_64(struct kvm_vcpu *vcpu, struct kvm_run *run) kvm_vcpu 2235 arch/arm64/kvm/sys_regs.c int kvm_handle_cp14_32(struct kvm_vcpu *vcpu, struct kvm_run *run) kvm_vcpu 2242 arch/arm64/kvm/sys_regs.c static int emulate_sys_reg(struct kvm_vcpu *vcpu, kvm_vcpu 2266 arch/arm64/kvm/sys_regs.c static void reset_sys_reg_descs(struct kvm_vcpu *vcpu, kvm_vcpu 2287 arch/arm64/kvm/sys_regs.c int kvm_handle_sys_reg(struct kvm_vcpu *vcpu, struct kvm_run *run) kvm_vcpu 2358 arch/arm64/kvm/sys_regs.c static const struct sys_reg_desc *index_to_sys_reg_desc(struct kvm_vcpu *vcpu, kvm_vcpu 2393 arch/arm64/kvm/sys_regs.c static void get_##reg(struct kvm_vcpu *v, \ kvm_vcpu 2404 arch/arm64/kvm/sys_regs.c static void get_ctr_el0(struct kvm_vcpu *v, const struct sys_reg_desc *r) kvm_vcpu 2550 arch/arm64/kvm/sys_regs.c int kvm_arm_sys_reg_get_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg) kvm_vcpu 2575 arch/arm64/kvm/sys_regs.c int kvm_arm_sys_reg_set_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg) kvm_vcpu 2650 arch/arm64/kvm/sys_regs.c static int walk_one_sys_reg(const struct kvm_vcpu *vcpu, kvm_vcpu 2673 arch/arm64/kvm/sys_regs.c static int walk_sys_regs(struct kvm_vcpu *vcpu, u64 __user *uind) kvm_vcpu 2708 arch/arm64/kvm/sys_regs.c unsigned long kvm_arm_num_sys_reg_descs(struct kvm_vcpu *vcpu) kvm_vcpu 2715 arch/arm64/kvm/sys_regs.c int kvm_arm_copy_sys_reg_indices(struct kvm_vcpu *vcpu, u64 __user *uindices) kvm_vcpu 2792 arch/arm64/kvm/sys_regs.c void kvm_reset_sys_regs(struct kvm_vcpu *vcpu) kvm_vcpu 38 arch/arm64/kvm/sys_regs.h bool (*access)(struct kvm_vcpu *, kvm_vcpu 43 arch/arm64/kvm/sys_regs.h void (*reset)(struct kvm_vcpu *, const struct sys_reg_desc *); kvm_vcpu 52 arch/arm64/kvm/sys_regs.h int (*get_user)(struct kvm_vcpu *vcpu, const struct sys_reg_desc *rd, kvm_vcpu 54 arch/arm64/kvm/sys_regs.h int (*set_user)(struct kvm_vcpu *vcpu, const struct sys_reg_desc *rd, kvm_vcpu 58 arch/arm64/kvm/sys_regs.h unsigned int (*visibility)(const struct kvm_vcpu *vcpu, kvm_vcpu 72 arch/arm64/kvm/sys_regs.h static inline bool ignore_write(struct kvm_vcpu *vcpu, kvm_vcpu 78 arch/arm64/kvm/sys_regs.h static inline bool read_zero(struct kvm_vcpu *vcpu, kvm_vcpu 86 arch/arm64/kvm/sys_regs.h static inline void reset_unknown(struct kvm_vcpu *vcpu, kvm_vcpu 94 arch/arm64/kvm/sys_regs.h static inline void reset_val(struct kvm_vcpu *vcpu, const struct sys_reg_desc *r) kvm_vcpu 101 arch/arm64/kvm/sys_regs.h static inline bool sysreg_hidden_from_guest(const struct kvm_vcpu *vcpu, kvm_vcpu 110 arch/arm64/kvm/sys_regs.h static inline bool sysreg_hidden_from_user(const struct kvm_vcpu *vcpu, kvm_vcpu 23 arch/arm64/kvm/sys_regs_generic_v8.c static bool access_actlr(struct kvm_vcpu *vcpu, kvm_vcpu 34 arch/arm64/kvm/sys_regs_generic_v8.c static void reset_actlr(struct kvm_vcpu *vcpu, const struct sys_reg_desc *r) kvm_vcpu 50 arch/arm64/kvm/trace.h TP_PROTO(struct kvm_vcpu *vcpu, __u32 guest_debug), kvm_vcpu 54 arch/arm64/kvm/trace.h __field(struct kvm_vcpu *, vcpu) kvm_vcpu 191 arch/arm64/kvm/trace.h TP_PROTO(struct kvm_vcpu *vcpu, __u32 guest_debug), kvm_vcpu 195 arch/arm64/kvm/trace.h __field(struct kvm_vcpu *, vcpu) kvm_vcpu 13 arch/arm64/kvm/vgic-sys-reg-v3.c static bool access_gic_ctlr(struct kvm_vcpu *vcpu, struct sys_reg_params *p, kvm_vcpu 87 arch/arm64/kvm/vgic-sys-reg-v3.c static bool access_gic_pmr(struct kvm_vcpu *vcpu, struct sys_reg_params *p, kvm_vcpu 103 arch/arm64/kvm/vgic-sys-reg-v3.c static bool access_gic_bpr0(struct kvm_vcpu *vcpu, struct sys_reg_params *p, kvm_vcpu 121 arch/arm64/kvm/vgic-sys-reg-v3.c static bool access_gic_bpr1(struct kvm_vcpu *vcpu, struct sys_reg_params *p, kvm_vcpu 147 arch/arm64/kvm/vgic-sys-reg-v3.c static bool access_gic_grpen0(struct kvm_vcpu *vcpu, struct sys_reg_params *p, kvm_vcpu 165 arch/arm64/kvm/vgic-sys-reg-v3.c static bool access_gic_grpen1(struct kvm_vcpu *vcpu, struct sys_reg_params *p, kvm_vcpu 183 arch/arm64/kvm/vgic-sys-reg-v3.c static void vgic_v3_access_apr_reg(struct kvm_vcpu *vcpu, kvm_vcpu 200 arch/arm64/kvm/vgic-sys-reg-v3.c static bool access_gic_aprn(struct kvm_vcpu *vcpu, struct sys_reg_params *p, kvm_vcpu 217 arch/arm64/kvm/vgic-sys-reg-v3.c static bool access_gic_ap0r(struct kvm_vcpu *vcpu, struct sys_reg_params *p, kvm_vcpu 224 arch/arm64/kvm/vgic-sys-reg-v3.c static bool access_gic_ap1r(struct kvm_vcpu *vcpu, struct sys_reg_params *p, kvm_vcpu 230 arch/arm64/kvm/vgic-sys-reg-v3.c static bool access_gic_sre(struct kvm_vcpu *vcpu, struct sys_reg_params *p, kvm_vcpu 263 arch/arm64/kvm/vgic-sys-reg-v3.c int vgic_v3_has_cpu_sysregs_attr(struct kvm_vcpu *vcpu, bool is_write, u64 id, kvm_vcpu 281 arch/arm64/kvm/vgic-sys-reg-v3.c int vgic_v3_cpu_sysregs_uaccess(struct kvm_vcpu *vcpu, bool is_write, u64 id, kvm_vcpu 317 arch/mips/include/asm/kvm_host.h int (*vcpu_run)(struct kvm_run *run, struct kvm_vcpu *vcpu); kvm_vcpu 772 arch/mips/include/asm/kvm_host.h int (*handle_cop_unusable)(struct kvm_vcpu *vcpu); kvm_vcpu 773 arch/mips/include/asm/kvm_host.h int (*handle_tlb_mod)(struct kvm_vcpu *vcpu); kvm_vcpu 774 arch/mips/include/asm/kvm_host.h int (*handle_tlb_ld_miss)(struct kvm_vcpu *vcpu); kvm_vcpu 775 arch/mips/include/asm/kvm_host.h int (*handle_tlb_st_miss)(struct kvm_vcpu *vcpu); kvm_vcpu 776 arch/mips/include/asm/kvm_host.h int (*handle_addr_err_st)(struct kvm_vcpu *vcpu); kvm_vcpu 777 arch/mips/include/asm/kvm_host.h int (*handle_addr_err_ld)(struct kvm_vcpu *vcpu); kvm_vcpu 778 arch/mips/include/asm/kvm_host.h int (*handle_syscall)(struct kvm_vcpu *vcpu); kvm_vcpu 779 arch/mips/include/asm/kvm_host.h int (*handle_res_inst)(struct kvm_vcpu *vcpu); kvm_vcpu 780 arch/mips/include/asm/kvm_host.h int (*handle_break)(struct kvm_vcpu *vcpu); kvm_vcpu 781 arch/mips/include/asm/kvm_host.h int (*handle_trap)(struct kvm_vcpu *vcpu); kvm_vcpu 782 arch/mips/include/asm/kvm_host.h int (*handle_msa_fpe)(struct kvm_vcpu *vcpu); kvm_vcpu 783 arch/mips/include/asm/kvm_host.h int (*handle_fpe)(struct kvm_vcpu *vcpu); kvm_vcpu 784 arch/mips/include/asm/kvm_host.h int (*handle_msa_disabled)(struct kvm_vcpu *vcpu); kvm_vcpu 785 arch/mips/include/asm/kvm_host.h int (*handle_guest_exit)(struct kvm_vcpu *vcpu); kvm_vcpu 789 arch/mips/include/asm/kvm_host.h int (*vcpu_init)(struct kvm_vcpu *vcpu); kvm_vcpu 790 arch/mips/include/asm/kvm_host.h void (*vcpu_uninit)(struct kvm_vcpu *vcpu); kvm_vcpu 791 arch/mips/include/asm/kvm_host.h int (*vcpu_setup)(struct kvm_vcpu *vcpu); kvm_vcpu 801 arch/mips/include/asm/kvm_host.h void (*queue_timer_int)(struct kvm_vcpu *vcpu); kvm_vcpu 802 arch/mips/include/asm/kvm_host.h void (*dequeue_timer_int)(struct kvm_vcpu *vcpu); kvm_vcpu 803 arch/mips/include/asm/kvm_host.h void (*queue_io_int)(struct kvm_vcpu *vcpu, kvm_vcpu 805 arch/mips/include/asm/kvm_host.h void (*dequeue_io_int)(struct kvm_vcpu *vcpu, kvm_vcpu 807 arch/mips/include/asm/kvm_host.h int (*irq_deliver)(struct kvm_vcpu *vcpu, unsigned int priority, kvm_vcpu 809 arch/mips/include/asm/kvm_host.h int (*irq_clear)(struct kvm_vcpu *vcpu, unsigned int priority, kvm_vcpu 811 arch/mips/include/asm/kvm_host.h unsigned long (*num_regs)(struct kvm_vcpu *vcpu); kvm_vcpu 812 arch/mips/include/asm/kvm_host.h int (*copy_reg_indices)(struct kvm_vcpu *vcpu, u64 __user *indices); kvm_vcpu 813 arch/mips/include/asm/kvm_host.h int (*get_one_reg)(struct kvm_vcpu *vcpu, kvm_vcpu 815 arch/mips/include/asm/kvm_host.h int (*set_one_reg)(struct kvm_vcpu *vcpu, kvm_vcpu 817 arch/mips/include/asm/kvm_host.h int (*vcpu_load)(struct kvm_vcpu *vcpu, int cpu); kvm_vcpu 818 arch/mips/include/asm/kvm_host.h int (*vcpu_put)(struct kvm_vcpu *vcpu, int cpu); kvm_vcpu 819 arch/mips/include/asm/kvm_host.h int (*vcpu_run)(struct kvm_run *run, struct kvm_vcpu *vcpu); kvm_vcpu 820 arch/mips/include/asm/kvm_host.h void (*vcpu_reenter)(struct kvm_run *run, struct kvm_vcpu *vcpu); kvm_vcpu 826 arch/mips/include/asm/kvm_host.h int kvm_arch_vcpu_dump_regs(struct kvm_vcpu *vcpu); kvm_vcpu 828 arch/mips/include/asm/kvm_host.h extern int kvm_mips_handle_exit(struct kvm_run *run, struct kvm_vcpu *vcpu); kvm_vcpu 845 arch/mips/include/asm/kvm_host.h void kvm_own_fpu(struct kvm_vcpu *vcpu); kvm_vcpu 846 arch/mips/include/asm/kvm_host.h void kvm_own_msa(struct kvm_vcpu *vcpu); kvm_vcpu 847 arch/mips/include/asm/kvm_host.h void kvm_drop_fpu(struct kvm_vcpu *vcpu); kvm_vcpu 848 arch/mips/include/asm/kvm_host.h void kvm_lose_fpu(struct kvm_vcpu *vcpu); kvm_vcpu 851 arch/mips/include/asm/kvm_host.h u32 kvm_get_kernel_asid(struct kvm_vcpu *vcpu); kvm_vcpu 853 arch/mips/include/asm/kvm_host.h u32 kvm_get_user_asid(struct kvm_vcpu *vcpu); kvm_vcpu 855 arch/mips/include/asm/kvm_host.h u32 kvm_get_commpage_asid (struct kvm_vcpu *vcpu); kvm_vcpu 859 arch/mips/include/asm/kvm_host.h struct kvm_vcpu *vcpu, bool write_fault); kvm_vcpu 862 arch/mips/include/asm/kvm_host.h struct kvm_vcpu *vcpu, kvm_vcpu 866 arch/mips/include/asm/kvm_host.h struct kvm_vcpu *vcpu); kvm_vcpu 868 arch/mips/include/asm/kvm_host.h extern int kvm_mips_handle_mapped_seg_tlb_fault(struct kvm_vcpu *vcpu, kvm_vcpu 876 arch/mips/include/asm/kvm_host.h struct kvm_vcpu *vcpu, kvm_vcpu 880 arch/mips/include/asm/kvm_host.h extern void kvm_mips_dump_guest_tlbs(struct kvm_vcpu *vcpu); kvm_vcpu 881 arch/mips/include/asm/kvm_host.h extern int kvm_mips_host_tlb_inv(struct kvm_vcpu *vcpu, unsigned long entryhi, kvm_vcpu 884 arch/mips/include/asm/kvm_host.h extern int kvm_mips_guest_tlb_lookup(struct kvm_vcpu *vcpu, kvm_vcpu 888 arch/mips/include/asm/kvm_host.h int kvm_vz_host_tlb_inv(struct kvm_vcpu *vcpu, unsigned long entryhi); kvm_vcpu 889 arch/mips/include/asm/kvm_host.h int kvm_vz_guest_tlb_lookup(struct kvm_vcpu *vcpu, unsigned long gva, kvm_vcpu 922 arch/mips/include/asm/kvm_host.h void kvm_mmu_free_memory_caches(struct kvm_vcpu *vcpu); kvm_vcpu 923 arch/mips/include/asm/kvm_host.h void kvm_trap_emul_invalidate_gva(struct kvm_vcpu *vcpu, unsigned long addr, kvm_vcpu 925 arch/mips/include/asm/kvm_host.h void kvm_trap_emul_gva_lockless_begin(struct kvm_vcpu *vcpu); kvm_vcpu 926 arch/mips/include/asm/kvm_host.h void kvm_trap_emul_gva_lockless_end(struct kvm_vcpu *vcpu); kvm_vcpu 936 arch/mips/include/asm/kvm_host.h enum kvm_mips_fault_result kvm_trap_emul_gva_fault(struct kvm_vcpu *vcpu, kvm_vcpu 948 arch/mips/include/asm/kvm_host.h int kvm_get_inst(u32 *opc, struct kvm_vcpu *vcpu, u32 *out); kvm_vcpu 949 arch/mips/include/asm/kvm_host.h enum emulation_result update_pc(struct kvm_vcpu *vcpu, u32 cause); kvm_vcpu 950 arch/mips/include/asm/kvm_host.h int kvm_get_badinstr(u32 *opc, struct kvm_vcpu *vcpu, u32 *out); kvm_vcpu 951 arch/mips/include/asm/kvm_host.h int kvm_get_badinstrp(u32 *opc, struct kvm_vcpu *vcpu, u32 *out); kvm_vcpu 983 arch/mips/include/asm/kvm_host.h struct kvm_vcpu *vcpu); kvm_vcpu 985 arch/mips/include/asm/kvm_host.h long kvm_mips_guest_exception_base(struct kvm_vcpu *vcpu); kvm_vcpu 990 arch/mips/include/asm/kvm_host.h struct kvm_vcpu *vcpu); kvm_vcpu 995 arch/mips/include/asm/kvm_host.h struct kvm_vcpu *vcpu); kvm_vcpu 1000 arch/mips/include/asm/kvm_host.h struct kvm_vcpu *vcpu); kvm_vcpu 1005 arch/mips/include/asm/kvm_host.h struct kvm_vcpu *vcpu); kvm_vcpu 1010 arch/mips/include/asm/kvm_host.h struct kvm_vcpu *vcpu); kvm_vcpu 1015 arch/mips/include/asm/kvm_host.h struct kvm_vcpu *vcpu); kvm_vcpu 1020 arch/mips/include/asm/kvm_host.h struct kvm_vcpu *vcpu); kvm_vcpu 1025 arch/mips/include/asm/kvm_host.h struct kvm_vcpu *vcpu); kvm_vcpu 1030 arch/mips/include/asm/kvm_host.h struct kvm_vcpu *vcpu); kvm_vcpu 1035 arch/mips/include/asm/kvm_host.h struct kvm_vcpu *vcpu); kvm_vcpu 1040 arch/mips/include/asm/kvm_host.h struct kvm_vcpu *vcpu); kvm_vcpu 1045 arch/mips/include/asm/kvm_host.h struct kvm_vcpu *vcpu); kvm_vcpu 1050 arch/mips/include/asm/kvm_host.h struct kvm_vcpu *vcpu); kvm_vcpu 1055 arch/mips/include/asm/kvm_host.h struct kvm_vcpu *vcpu); kvm_vcpu 1057 arch/mips/include/asm/kvm_host.h extern enum emulation_result kvm_mips_complete_mmio_load(struct kvm_vcpu *vcpu, kvm_vcpu 1060 arch/mips/include/asm/kvm_host.h u32 kvm_mips_read_count(struct kvm_vcpu *vcpu); kvm_vcpu 1061 arch/mips/include/asm/kvm_host.h void kvm_mips_write_count(struct kvm_vcpu *vcpu, u32 count); kvm_vcpu 1062 arch/mips/include/asm/kvm_host.h void kvm_mips_write_compare(struct kvm_vcpu *vcpu, u32 compare, bool ack); kvm_vcpu 1063 arch/mips/include/asm/kvm_host.h void kvm_mips_init_count(struct kvm_vcpu *vcpu, unsigned long count_hz); kvm_vcpu 1064 arch/mips/include/asm/kvm_host.h int kvm_mips_set_count_ctl(struct kvm_vcpu *vcpu, s64 count_ctl); kvm_vcpu 1065 arch/mips/include/asm/kvm_host.h int kvm_mips_set_count_resume(struct kvm_vcpu *vcpu, s64 count_resume); kvm_vcpu 1066 arch/mips/include/asm/kvm_host.h int kvm_mips_set_count_hz(struct kvm_vcpu *vcpu, s64 count_hz); kvm_vcpu 1067 arch/mips/include/asm/kvm_host.h void kvm_mips_count_enable_cause(struct kvm_vcpu *vcpu); kvm_vcpu 1068 arch/mips/include/asm/kvm_host.h void kvm_mips_count_disable_cause(struct kvm_vcpu *vcpu); kvm_vcpu 1069 arch/mips/include/asm/kvm_host.h enum hrtimer_restart kvm_mips_count_timeout(struct kvm_vcpu *vcpu); kvm_vcpu 1072 arch/mips/include/asm/kvm_host.h int kvm_mips_count_disabled(struct kvm_vcpu *vcpu); kvm_vcpu 1073 arch/mips/include/asm/kvm_host.h ktime_t kvm_mips_freeze_hrtimer(struct kvm_vcpu *vcpu, u32 *count); kvm_vcpu 1074 arch/mips/include/asm/kvm_host.h int kvm_mips_restore_hrtimer(struct kvm_vcpu *vcpu, ktime_t before, kvm_vcpu 1078 arch/mips/include/asm/kvm_host.h void kvm_vz_acquire_htimer(struct kvm_vcpu *vcpu); kvm_vcpu 1079 arch/mips/include/asm/kvm_host.h void kvm_vz_lose_htimer(struct kvm_vcpu *vcpu); kvm_vcpu 1081 arch/mips/include/asm/kvm_host.h static inline void kvm_vz_acquire_htimer(struct kvm_vcpu *vcpu) {} kvm_vcpu 1082 arch/mips/include/asm/kvm_host.h static inline void kvm_vz_lose_htimer(struct kvm_vcpu *vcpu) {} kvm_vcpu 1088 arch/mips/include/asm/kvm_host.h struct kvm_vcpu *vcpu); kvm_vcpu 1094 arch/mips/include/asm/kvm_host.h struct kvm_vcpu *vcpu); kvm_vcpu 1099 arch/mips/include/asm/kvm_host.h struct kvm_vcpu *vcpu); kvm_vcpu 1103 arch/mips/include/asm/kvm_host.h struct kvm_vcpu *vcpu); kvm_vcpu 1107 arch/mips/include/asm/kvm_host.h struct kvm_vcpu *vcpu); kvm_vcpu 1110 arch/mips/include/asm/kvm_host.h enum emulation_result kvm_mips_emul_wait(struct kvm_vcpu *vcpu); kvm_vcpu 1112 arch/mips/include/asm/kvm_host.h unsigned int kvm_mips_config1_wrmask(struct kvm_vcpu *vcpu); kvm_vcpu 1113 arch/mips/include/asm/kvm_host.h unsigned int kvm_mips_config3_wrmask(struct kvm_vcpu *vcpu); kvm_vcpu 1114 arch/mips/include/asm/kvm_host.h unsigned int kvm_mips_config4_wrmask(struct kvm_vcpu *vcpu); kvm_vcpu 1115 arch/mips/include/asm/kvm_host.h unsigned int kvm_mips_config5_wrmask(struct kvm_vcpu *vcpu); kvm_vcpu 1119 arch/mips/include/asm/kvm_host.h enum emulation_result kvm_mips_emul_hypcall(struct kvm_vcpu *vcpu, kvm_vcpu 1121 arch/mips/include/asm/kvm_host.h int kvm_mips_handle_hypcall(struct kvm_vcpu *vcpu); kvm_vcpu 1125 arch/mips/include/asm/kvm_host.h u32 *opc, struct kvm_vcpu *vcpu); kvm_vcpu 1127 arch/mips/include/asm/kvm_host.h struct kvm_vcpu *vcpu); kvm_vcpu 1129 arch/mips/include/asm/kvm_host.h struct kvm_vcpu *vcpu); kvm_vcpu 1131 arch/mips/include/asm/kvm_host.h struct kvm_vcpu *vcpu); kvm_vcpu 1134 arch/mips/include/asm/kvm_host.h extern void kvm_mips_dump_stats(struct kvm_vcpu *vcpu); kvm_vcpu 1142 arch/mips/include/asm/kvm_host.h static inline void kvm_arch_sched_in(struct kvm_vcpu *vcpu, int cpu) {} kvm_vcpu 1143 arch/mips/include/asm/kvm_host.h static inline void kvm_arch_vcpu_blocking(struct kvm_vcpu *vcpu) {} kvm_vcpu 1144 arch/mips/include/asm/kvm_host.h static inline void kvm_arch_vcpu_unblocking(struct kvm_vcpu *vcpu) {} kvm_vcpu 1145 arch/mips/include/asm/kvm_host.h static inline void kvm_arch_vcpu_block_finish(struct kvm_vcpu *vcpu) {} kvm_vcpu 26 arch/mips/kvm/commpage.c void kvm_mips_commpage_init(struct kvm_vcpu *vcpu) kvm_vcpu 22 arch/mips/kvm/commpage.h extern void kvm_mips_commpage_init(struct kvm_vcpu *vcpu); kvm_vcpu 30 arch/mips/kvm/dyntrans.c static int kvm_mips_trans_replace(struct kvm_vcpu *vcpu, u32 *opc, kvm_vcpu 69 arch/mips/kvm/dyntrans.c struct kvm_vcpu *vcpu) kvm_vcpu 82 arch/mips/kvm/dyntrans.c struct kvm_vcpu *vcpu) kvm_vcpu 98 arch/mips/kvm/dyntrans.c struct kvm_vcpu *vcpu) kvm_vcpu 125 arch/mips/kvm/dyntrans.c struct kvm_vcpu *vcpu) kvm_vcpu 41 arch/mips/kvm/emulate.c static int kvm_compute_return_epc(struct kvm_vcpu *vcpu, unsigned long instpc, kvm_vcpu 244 arch/mips/kvm/emulate.c enum emulation_result update_pc(struct kvm_vcpu *vcpu, u32 cause) kvm_vcpu 273 arch/mips/kvm/emulate.c int kvm_get_badinstr(u32 *opc, struct kvm_vcpu *vcpu, u32 *out) kvm_vcpu 294 arch/mips/kvm/emulate.c int kvm_get_badinstrp(u32 *opc, struct kvm_vcpu *vcpu, u32 *out) kvm_vcpu 312 arch/mips/kvm/emulate.c int kvm_mips_count_disabled(struct kvm_vcpu *vcpu) kvm_vcpu 327 arch/mips/kvm/emulate.c static u32 kvm_mips_ktime_to_count(struct kvm_vcpu *vcpu, ktime_t now) kvm_vcpu 366 arch/mips/kvm/emulate.c static inline ktime_t kvm_mips_count_time(struct kvm_vcpu *vcpu) kvm_vcpu 384 arch/mips/kvm/emulate.c static u32 kvm_mips_read_count_running(struct kvm_vcpu *vcpu, ktime_t now) kvm_vcpu 444 arch/mips/kvm/emulate.c u32 kvm_mips_read_count(struct kvm_vcpu *vcpu) kvm_vcpu 471 arch/mips/kvm/emulate.c ktime_t kvm_mips_freeze_hrtimer(struct kvm_vcpu *vcpu, u32 *count) kvm_vcpu 501 arch/mips/kvm/emulate.c static void kvm_mips_resume_hrtimer(struct kvm_vcpu *vcpu, kvm_vcpu 542 arch/mips/kvm/emulate.c int kvm_mips_restore_hrtimer(struct kvm_vcpu *vcpu, ktime_t before, kvm_vcpu 603 arch/mips/kvm/emulate.c void kvm_mips_write_count(struct kvm_vcpu *vcpu, u32 count) kvm_vcpu 628 arch/mips/kvm/emulate.c void kvm_mips_init_count(struct kvm_vcpu *vcpu, unsigned long count_hz) kvm_vcpu 649 arch/mips/kvm/emulate.c int kvm_mips_set_count_hz(struct kvm_vcpu *vcpu, s64 count_hz) kvm_vcpu 696 arch/mips/kvm/emulate.c void kvm_mips_write_compare(struct kvm_vcpu *vcpu, u32 compare, bool ack) kvm_vcpu 781 arch/mips/kvm/emulate.c static ktime_t kvm_mips_count_disable(struct kvm_vcpu *vcpu) kvm_vcpu 808 arch/mips/kvm/emulate.c void kvm_mips_count_disable_cause(struct kvm_vcpu *vcpu) kvm_vcpu 828 arch/mips/kvm/emulate.c void kvm_mips_count_enable_cause(struct kvm_vcpu *vcpu) kvm_vcpu 854 arch/mips/kvm/emulate.c int kvm_mips_set_count_ctl(struct kvm_vcpu *vcpu, s64 count_ctl) kvm_vcpu 916 arch/mips/kvm/emulate.c int kvm_mips_set_count_resume(struct kvm_vcpu *vcpu, s64 count_resume) kvm_vcpu 938 arch/mips/kvm/emulate.c enum hrtimer_restart kvm_mips_count_timeout(struct kvm_vcpu *vcpu) kvm_vcpu 946 arch/mips/kvm/emulate.c enum emulation_result kvm_mips_emul_eret(struct kvm_vcpu *vcpu) kvm_vcpu 969 arch/mips/kvm/emulate.c enum emulation_result kvm_mips_emul_wait(struct kvm_vcpu *vcpu) kvm_vcpu 994 arch/mips/kvm/emulate.c static void kvm_mips_change_entryhi(struct kvm_vcpu *vcpu, kvm_vcpu 1029 arch/mips/kvm/emulate.c enum emulation_result kvm_mips_emul_tlbr(struct kvm_vcpu *vcpu) kvm_vcpu 1060 arch/mips/kvm/emulate.c static void kvm_mips_invalidate_guest_tlb(struct kvm_vcpu *vcpu, kvm_vcpu 1102 arch/mips/kvm/emulate.c enum emulation_result kvm_mips_emul_tlbwi(struct kvm_vcpu *vcpu) kvm_vcpu 1138 arch/mips/kvm/emulate.c enum emulation_result kvm_mips_emul_tlbwr(struct kvm_vcpu *vcpu) kvm_vcpu 1163 arch/mips/kvm/emulate.c enum emulation_result kvm_mips_emul_tlbp(struct kvm_vcpu *vcpu) kvm_vcpu 1187 arch/mips/kvm/emulate.c unsigned int kvm_mips_config1_wrmask(struct kvm_vcpu *vcpu) kvm_vcpu 1205 arch/mips/kvm/emulate.c unsigned int kvm_mips_config3_wrmask(struct kvm_vcpu *vcpu) kvm_vcpu 1224 arch/mips/kvm/emulate.c unsigned int kvm_mips_config4_wrmask(struct kvm_vcpu *vcpu) kvm_vcpu 1242 arch/mips/kvm/emulate.c unsigned int kvm_mips_config5_wrmask(struct kvm_vcpu *vcpu) kvm_vcpu 1266 arch/mips/kvm/emulate.c struct kvm_vcpu *vcpu) kvm_vcpu 1601 arch/mips/kvm/emulate.c struct kvm_vcpu *vcpu) kvm_vcpu 1682 arch/mips/kvm/emulate.c struct kvm_vcpu *vcpu) kvm_vcpu 1756 arch/mips/kvm/emulate.c struct kvm_vcpu *vcpu, kvm_vcpu 1800 arch/mips/kvm/emulate.c struct kvm_vcpu *vcpu) kvm_vcpu 1933 arch/mips/kvm/emulate.c struct kvm_vcpu *vcpu) kvm_vcpu 1991 arch/mips/kvm/emulate.c long kvm_mips_guest_exception_base(struct kvm_vcpu *vcpu) kvm_vcpu 2004 arch/mips/kvm/emulate.c struct kvm_vcpu *vcpu) kvm_vcpu 2039 arch/mips/kvm/emulate.c struct kvm_vcpu *vcpu) kvm_vcpu 2083 arch/mips/kvm/emulate.c struct kvm_vcpu *vcpu) kvm_vcpu 2125 arch/mips/kvm/emulate.c struct kvm_vcpu *vcpu) kvm_vcpu 2167 arch/mips/kvm/emulate.c struct kvm_vcpu *vcpu) kvm_vcpu 2208 arch/mips/kvm/emulate.c struct kvm_vcpu *vcpu) kvm_vcpu 2248 arch/mips/kvm/emulate.c struct kvm_vcpu *vcpu) kvm_vcpu 2277 arch/mips/kvm/emulate.c struct kvm_vcpu *vcpu) kvm_vcpu 2312 arch/mips/kvm/emulate.c struct kvm_vcpu *vcpu) kvm_vcpu 2347 arch/mips/kvm/emulate.c struct kvm_vcpu *vcpu) kvm_vcpu 2382 arch/mips/kvm/emulate.c struct kvm_vcpu *vcpu) kvm_vcpu 2417 arch/mips/kvm/emulate.c struct kvm_vcpu *vcpu) kvm_vcpu 2452 arch/mips/kvm/emulate.c struct kvm_vcpu *vcpu) kvm_vcpu 2486 arch/mips/kvm/emulate.c struct kvm_vcpu *vcpu) kvm_vcpu 2577 arch/mips/kvm/emulate.c enum emulation_result kvm_mips_complete_mmio_load(struct kvm_vcpu *vcpu, kvm_vcpu 2626 arch/mips/kvm/emulate.c struct kvm_vcpu *vcpu) kvm_vcpu 2664 arch/mips/kvm/emulate.c struct kvm_vcpu *vcpu) kvm_vcpu 2760 arch/mips/kvm/emulate.c struct kvm_vcpu *vcpu, kvm_vcpu 243 arch/mips/kvm/entry.c UASM_i_ADDIU(&p, K1, A1, offsetof(struct kvm_vcpu, arch)); kvm_vcpu 321 arch/mips/kvm/entry.c UASM_i_LW(&p, S0, (int)offsetof(struct kvm_vcpu, kvm) - kvm_vcpu 322 arch/mips/kvm/entry.c (int)offsetof(struct kvm_vcpu, arch), K1); kvm_vcpu 485 arch/mips/kvm/entry.c UASM_i_SW(&p, K0, offsetof(struct kvm_vcpu, arch.gprs[K0]), K1); kvm_vcpu 522 arch/mips/kvm/entry.c UASM_i_LW(&p, K0, offsetof(struct kvm_vcpu, arch.gprs[K0]), K1); kvm_vcpu 558 arch/mips/kvm/entry.c UASM_i_ADDIU(&p, K1, K1, offsetof(struct kvm_vcpu, arch)); kvm_vcpu 634 arch/mips/kvm/entry.c UASM_i_LW(&p, S0, offsetof(struct kvm_vcpu, run), S1); kvm_vcpu 839 arch/mips/kvm/entry.c UASM_i_ADDIU(&p, K1, K1, offsetof(struct kvm_vcpu, arch)); kvm_vcpu 17 arch/mips/kvm/hypcall.c enum emulation_result kvm_mips_emul_hypcall(struct kvm_vcpu *vcpu, kvm_vcpu 32 arch/mips/kvm/hypcall.c static int kvm_mips_hypercall(struct kvm_vcpu *vcpu, unsigned long num, kvm_vcpu 40 arch/mips/kvm/hypcall.c int kvm_mips_handle_hypcall(struct kvm_vcpu *vcpu) kvm_vcpu 24 arch/mips/kvm/interrupt.c void kvm_mips_queue_irq(struct kvm_vcpu *vcpu, unsigned int priority) kvm_vcpu 29 arch/mips/kvm/interrupt.c void kvm_mips_dequeue_irq(struct kvm_vcpu *vcpu, unsigned int priority) kvm_vcpu 34 arch/mips/kvm/interrupt.c void kvm_mips_queue_timer_int_cb(struct kvm_vcpu *vcpu) kvm_vcpu 48 arch/mips/kvm/interrupt.c void kvm_mips_dequeue_timer_int_cb(struct kvm_vcpu *vcpu) kvm_vcpu 54 arch/mips/kvm/interrupt.c void kvm_mips_queue_io_int_cb(struct kvm_vcpu *vcpu, kvm_vcpu 87 arch/mips/kvm/interrupt.c void kvm_mips_dequeue_io_int_cb(struct kvm_vcpu *vcpu, kvm_vcpu 115 arch/mips/kvm/interrupt.c int kvm_mips_irq_deliver_cb(struct kvm_vcpu *vcpu, unsigned int priority, kvm_vcpu 198 arch/mips/kvm/interrupt.c int kvm_mips_irq_clear_cb(struct kvm_vcpu *vcpu, unsigned int priority, kvm_vcpu 204 arch/mips/kvm/interrupt.c void kvm_mips_deliver_interrupts(struct kvm_vcpu *vcpu, u32 cause) kvm_vcpu 239 arch/mips/kvm/interrupt.c int kvm_mips_pending_timer(struct kvm_vcpu *vcpu) kvm_vcpu 41 arch/mips/kvm/interrupt.h void kvm_mips_queue_irq(struct kvm_vcpu *vcpu, unsigned int priority); kvm_vcpu 42 arch/mips/kvm/interrupt.h void kvm_mips_dequeue_irq(struct kvm_vcpu *vcpu, unsigned int priority); kvm_vcpu 43 arch/mips/kvm/interrupt.h int kvm_mips_pending_timer(struct kvm_vcpu *vcpu); kvm_vcpu 45 arch/mips/kvm/interrupt.h void kvm_mips_queue_timer_int_cb(struct kvm_vcpu *vcpu); kvm_vcpu 46 arch/mips/kvm/interrupt.h void kvm_mips_dequeue_timer_int_cb(struct kvm_vcpu *vcpu); kvm_vcpu 47 arch/mips/kvm/interrupt.h void kvm_mips_queue_io_int_cb(struct kvm_vcpu *vcpu, kvm_vcpu 49 arch/mips/kvm/interrupt.h void kvm_mips_dequeue_io_int_cb(struct kvm_vcpu *vcpu, kvm_vcpu 51 arch/mips/kvm/interrupt.h int kvm_mips_irq_deliver_cb(struct kvm_vcpu *vcpu, unsigned int priority, kvm_vcpu 53 arch/mips/kvm/interrupt.h int kvm_mips_irq_clear_cb(struct kvm_vcpu *vcpu, unsigned int priority, kvm_vcpu 55 arch/mips/kvm/interrupt.h void kvm_mips_deliver_interrupts(struct kvm_vcpu *vcpu, u32 cause); kvm_vcpu 42 arch/mips/kvm/mips.c #define VCPU_STAT(x) offsetof(struct kvm_vcpu, stat.x) kvm_vcpu 96 arch/mips/kvm/mips.c int kvm_arch_vcpu_runnable(struct kvm_vcpu *vcpu) kvm_vcpu 101 arch/mips/kvm/mips.c bool kvm_arch_vcpu_in_kernel(struct kvm_vcpu *vcpu) kvm_vcpu 106 arch/mips/kvm/mips.c int kvm_arch_vcpu_should_kick(struct kvm_vcpu *vcpu) kvm_vcpu 156 arch/mips/kvm/mips.c struct kvm_vcpu *vcpu; kvm_vcpu 283 arch/mips/kvm/mips.c struct kvm_vcpu *kvm_arch_vcpu_create(struct kvm *kvm, unsigned int id) kvm_vcpu 289 arch/mips/kvm/mips.c struct kvm_vcpu *vcpu = kzalloc(sizeof(struct kvm_vcpu), GFP_KERNEL); kvm_vcpu 410 arch/mips/kvm/mips.c void kvm_arch_vcpu_free(struct kvm_vcpu *vcpu) kvm_vcpu 424 arch/mips/kvm/mips.c void kvm_arch_vcpu_destroy(struct kvm_vcpu *vcpu) kvm_vcpu 429 arch/mips/kvm/mips.c int kvm_arch_vcpu_ioctl_set_guest_debug(struct kvm_vcpu *vcpu, kvm_vcpu 435 arch/mips/kvm/mips.c int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu, struct kvm_run *run) kvm_vcpu 479 arch/mips/kvm/mips.c int kvm_vcpu_ioctl_interrupt(struct kvm_vcpu *vcpu, kvm_vcpu 483 arch/mips/kvm/mips.c struct kvm_vcpu *dvcpu = NULL; kvm_vcpu 513 arch/mips/kvm/mips.c int kvm_arch_vcpu_ioctl_get_mpstate(struct kvm_vcpu *vcpu, kvm_vcpu 519 arch/mips/kvm/mips.c int kvm_arch_vcpu_ioctl_set_mpstate(struct kvm_vcpu *vcpu, kvm_vcpu 576 arch/mips/kvm/mips.c static unsigned long kvm_mips_num_regs(struct kvm_vcpu *vcpu) kvm_vcpu 594 arch/mips/kvm/mips.c static int kvm_mips_copy_reg_indices(struct kvm_vcpu *vcpu, u64 __user *indices) kvm_vcpu 644 arch/mips/kvm/mips.c static int kvm_mips_get_reg(struct kvm_vcpu *vcpu, kvm_vcpu 756 arch/mips/kvm/mips.c static int kvm_mips_set_reg(struct kvm_vcpu *vcpu, kvm_vcpu 869 arch/mips/kvm/mips.c static int kvm_vcpu_ioctl_enable_cap(struct kvm_vcpu *vcpu, kvm_vcpu 899 arch/mips/kvm/mips.c struct kvm_vcpu *vcpu = filp->private_data; kvm_vcpu 919 arch/mips/kvm/mips.c struct kvm_vcpu *vcpu = filp->private_data; kvm_vcpu 1066 arch/mips/kvm/mips.c int kvm_arch_vcpu_ioctl_get_sregs(struct kvm_vcpu *vcpu, kvm_vcpu 1072 arch/mips/kvm/mips.c int kvm_arch_vcpu_ioctl_set_sregs(struct kvm_vcpu *vcpu, kvm_vcpu 1078 arch/mips/kvm/mips.c void kvm_arch_vcpu_postcreate(struct kvm_vcpu *vcpu) kvm_vcpu 1082 arch/mips/kvm/mips.c int kvm_arch_vcpu_ioctl_get_fpu(struct kvm_vcpu *vcpu, struct kvm_fpu *fpu) kvm_vcpu 1087 arch/mips/kvm/mips.c int kvm_arch_vcpu_ioctl_set_fpu(struct kvm_vcpu *vcpu, struct kvm_fpu *fpu) kvm_vcpu 1092 arch/mips/kvm/mips.c vm_fault_t kvm_arch_vcpu_fault(struct kvm_vcpu *vcpu, struct vm_fault *vmf) kvm_vcpu 1144 arch/mips/kvm/mips.c int kvm_cpu_has_pending_timer(struct kvm_vcpu *vcpu) kvm_vcpu 1150 arch/mips/kvm/mips.c int kvm_arch_vcpu_dump_regs(struct kvm_vcpu *vcpu) kvm_vcpu 1181 arch/mips/kvm/mips.c int kvm_arch_vcpu_ioctl_set_regs(struct kvm_vcpu *vcpu, struct kvm_regs *regs) kvm_vcpu 1198 arch/mips/kvm/mips.c int kvm_arch_vcpu_ioctl_get_regs(struct kvm_vcpu *vcpu, struct kvm_regs *regs) kvm_vcpu 1217 arch/mips/kvm/mips.c struct kvm_vcpu *vcpu = (struct kvm_vcpu *)data; kvm_vcpu 1229 arch/mips/kvm/mips.c struct kvm_vcpu *vcpu; kvm_vcpu 1231 arch/mips/kvm/mips.c vcpu = container_of(timer, struct kvm_vcpu, arch.comparecount_timer); kvm_vcpu 1236 arch/mips/kvm/mips.c int kvm_arch_vcpu_init(struct kvm_vcpu *vcpu) kvm_vcpu 1250 arch/mips/kvm/mips.c void kvm_arch_vcpu_uninit(struct kvm_vcpu *vcpu) kvm_vcpu 1255 arch/mips/kvm/mips.c int kvm_arch_vcpu_ioctl_translate(struct kvm_vcpu *vcpu, kvm_vcpu 1262 arch/mips/kvm/mips.c int kvm_arch_vcpu_setup(struct kvm_vcpu *vcpu) kvm_vcpu 1281 arch/mips/kvm/mips.c int kvm_mips_handle_exit(struct kvm_run *run, struct kvm_vcpu *vcpu) kvm_vcpu 1496 arch/mips/kvm/mips.c void kvm_own_fpu(struct kvm_vcpu *vcpu) kvm_vcpu 1544 arch/mips/kvm/mips.c void kvm_own_msa(struct kvm_vcpu *vcpu) kvm_vcpu 1606 arch/mips/kvm/mips.c void kvm_drop_fpu(struct kvm_vcpu *vcpu) kvm_vcpu 1623 arch/mips/kvm/mips.c void kvm_lose_fpu(struct kvm_vcpu *vcpu) kvm_vcpu 1728 arch/mips/kvm/mips.c ret = kvm_init(NULL, sizeof(struct kvm_vcpu), 0, THIS_MODULE); kvm_vcpu 60 arch/mips/kvm/mmu.c void kvm_mmu_free_memory_caches(struct kvm_vcpu *vcpu) kvm_vcpu 610 arch/mips/kvm/mmu.c static int _kvm_mips_map_page_fast(struct kvm_vcpu *vcpu, unsigned long gpa, kvm_vcpu 686 arch/mips/kvm/mmu.c static int kvm_mips_map_page(struct kvm_vcpu *vcpu, unsigned long gpa, kvm_vcpu 785 arch/mips/kvm/mmu.c static pte_t *kvm_trap_emul_pte_for_gva(struct kvm_vcpu *vcpu, kvm_vcpu 806 arch/mips/kvm/mmu.c void kvm_trap_emul_invalidate_gva(struct kvm_vcpu *vcpu, unsigned long addr, kvm_vcpu 988 arch/mips/kvm/mmu.c struct kvm_vcpu *vcpu, kvm_vcpu 1004 arch/mips/kvm/mmu.c struct kvm_vcpu *vcpu, kvm_vcpu 1040 arch/mips/kvm/mmu.c int kvm_mips_handle_mapped_seg_tlb_fault(struct kvm_vcpu *vcpu, kvm_vcpu 1098 arch/mips/kvm/mmu.c struct kvm_vcpu *vcpu) kvm_vcpu 1129 arch/mips/kvm/mmu.c static void kvm_mips_migrate_count(struct kvm_vcpu *vcpu) kvm_vcpu 1136 arch/mips/kvm/mmu.c void kvm_arch_vcpu_load(struct kvm_vcpu *vcpu, int cpu) kvm_vcpu 1163 arch/mips/kvm/mmu.c void kvm_arch_vcpu_put(struct kvm_vcpu *vcpu) kvm_vcpu 1196 arch/mips/kvm/mmu.c enum kvm_mips_fault_result kvm_trap_emul_gva_fault(struct kvm_vcpu *vcpu, kvm_vcpu 1231 arch/mips/kvm/mmu.c int kvm_get_inst(u32 *opc, struct kvm_vcpu *vcpu, u32 *out) kvm_vcpu 49 arch/mips/kvm/stats.c void kvm_mips_dump_stats(struct kvm_vcpu *vcpu) kvm_vcpu 44 arch/mips/kvm/tlb.c static u32 kvm_mips_get_root_asid(struct kvm_vcpu *vcpu) kvm_vcpu 55 arch/mips/kvm/tlb.c static u32 kvm_mips_get_kernel_asid(struct kvm_vcpu *vcpu) kvm_vcpu 63 arch/mips/kvm/tlb.c static u32 kvm_mips_get_user_asid(struct kvm_vcpu *vcpu) kvm_vcpu 88 arch/mips/kvm/tlb.c void kvm_mips_dump_guest_tlbs(struct kvm_vcpu *vcpu) kvm_vcpu 118 arch/mips/kvm/tlb.c int kvm_mips_guest_tlb_lookup(struct kvm_vcpu *vcpu, unsigned long entryhi) kvm_vcpu 166 arch/mips/kvm/tlb.c int kvm_mips_host_tlb_inv(struct kvm_vcpu *vcpu, unsigned long va, kvm_vcpu 251 arch/mips/kvm/tlb.c int kvm_vz_host_tlb_inv(struct kvm_vcpu *vcpu, unsigned long va) kvm_vcpu 304 arch/mips/kvm/tlb.c int kvm_vz_guest_tlb_lookup(struct kvm_vcpu *vcpu, unsigned long gva, kvm_vcpu 31 arch/mips/kvm/trace.h TP_PROTO(struct kvm_vcpu *vcpu), kvm_vcpu 46 arch/mips/kvm/trace.h TP_PROTO(struct kvm_vcpu *vcpu), kvm_vcpu 50 arch/mips/kvm/trace.h TP_PROTO(struct kvm_vcpu *vcpu), kvm_vcpu 54 arch/mips/kvm/trace.h TP_PROTO(struct kvm_vcpu *vcpu), kvm_vcpu 116 arch/mips/kvm/trace.h TP_PROTO(struct kvm_vcpu *vcpu, unsigned int reason), kvm_vcpu 203 arch/mips/kvm/trace.h TP_PROTO(struct kvm_vcpu *vcpu, unsigned int op, unsigned int reg, kvm_vcpu 253 arch/mips/kvm/trace.h TP_PROTO(struct kvm_vcpu *vcpu, unsigned int op, kvm_vcpu 277 arch/mips/kvm/trace.h TP_PROTO(struct kvm_vcpu *vcpu, unsigned int old_asid, kvm_vcpu 299 arch/mips/kvm/trace.h TP_PROTO(struct kvm_vcpu *vcpu, unsigned int guestid), kvm_vcpu 314 arch/mips/kvm/trace.h TP_PROTO(struct kvm_vcpu *vcpu), kvm_vcpu 44 arch/mips/kvm/trap_emul.c static int kvm_trap_emul_no_handler(struct kvm_vcpu *vcpu) kvm_vcpu 67 arch/mips/kvm/trap_emul.c static int kvm_trap_emul_handle_cop_unusable(struct kvm_vcpu *vcpu) kvm_vcpu 120 arch/mips/kvm/trap_emul.c struct kvm_vcpu *vcpu) kvm_vcpu 153 arch/mips/kvm/trap_emul.c struct kvm_vcpu *vcpu) kvm_vcpu 180 arch/mips/kvm/trap_emul.c struct kvm_vcpu *vcpu, bool store) kvm_vcpu 188 arch/mips/kvm/trap_emul.c static int kvm_trap_emul_handle_tlb_mod(struct kvm_vcpu *vcpu) kvm_vcpu 249 arch/mips/kvm/trap_emul.c static int kvm_trap_emul_handle_tlb_miss(struct kvm_vcpu *vcpu, bool store) kvm_vcpu 311 arch/mips/kvm/trap_emul.c static int kvm_trap_emul_handle_tlb_st_miss(struct kvm_vcpu *vcpu) kvm_vcpu 316 arch/mips/kvm/trap_emul.c static int kvm_trap_emul_handle_tlb_ld_miss(struct kvm_vcpu *vcpu) kvm_vcpu 321 arch/mips/kvm/trap_emul.c static int kvm_trap_emul_handle_addr_err_st(struct kvm_vcpu *vcpu) kvm_vcpu 341 arch/mips/kvm/trap_emul.c static int kvm_trap_emul_handle_addr_err_ld(struct kvm_vcpu *vcpu) kvm_vcpu 360 arch/mips/kvm/trap_emul.c static int kvm_trap_emul_handle_syscall(struct kvm_vcpu *vcpu) kvm_vcpu 378 arch/mips/kvm/trap_emul.c static int kvm_trap_emul_handle_res_inst(struct kvm_vcpu *vcpu) kvm_vcpu 396 arch/mips/kvm/trap_emul.c static int kvm_trap_emul_handle_break(struct kvm_vcpu *vcpu) kvm_vcpu 414 arch/mips/kvm/trap_emul.c static int kvm_trap_emul_handle_trap(struct kvm_vcpu *vcpu) kvm_vcpu 432 arch/mips/kvm/trap_emul.c static int kvm_trap_emul_handle_msa_fpe(struct kvm_vcpu *vcpu) kvm_vcpu 450 arch/mips/kvm/trap_emul.c static int kvm_trap_emul_handle_fpe(struct kvm_vcpu *vcpu) kvm_vcpu 474 arch/mips/kvm/trap_emul.c static int kvm_trap_emul_handle_msa_disabled(struct kvm_vcpu *vcpu) kvm_vcpu 540 arch/mips/kvm/trap_emul.c static int kvm_trap_emul_vcpu_init(struct kvm_vcpu *vcpu) kvm_vcpu 605 arch/mips/kvm/trap_emul.c static void kvm_trap_emul_vcpu_uninit(struct kvm_vcpu *vcpu) kvm_vcpu 611 arch/mips/kvm/trap_emul.c static int kvm_trap_emul_vcpu_setup(struct kvm_vcpu *vcpu) kvm_vcpu 749 arch/mips/kvm/trap_emul.c static unsigned long kvm_trap_emul_num_regs(struct kvm_vcpu *vcpu) kvm_vcpu 754 arch/mips/kvm/trap_emul.c static int kvm_trap_emul_copy_reg_indices(struct kvm_vcpu *vcpu, kvm_vcpu 765 arch/mips/kvm/trap_emul.c static int kvm_trap_emul_get_one_reg(struct kvm_vcpu *vcpu, kvm_vcpu 883 arch/mips/kvm/trap_emul.c static int kvm_trap_emul_set_one_reg(struct kvm_vcpu *vcpu, kvm_vcpu 1047 arch/mips/kvm/trap_emul.c static int kvm_trap_emul_vcpu_load(struct kvm_vcpu *vcpu, int cpu) kvm_vcpu 1067 arch/mips/kvm/trap_emul.c static int kvm_trap_emul_vcpu_put(struct kvm_vcpu *vcpu, int cpu) kvm_vcpu 1081 arch/mips/kvm/trap_emul.c static void kvm_trap_emul_check_requests(struct kvm_vcpu *vcpu, int cpu, kvm_vcpu 1127 arch/mips/kvm/trap_emul.c void kvm_trap_emul_gva_lockless_begin(struct kvm_vcpu *vcpu) kvm_vcpu 1166 arch/mips/kvm/trap_emul.c void kvm_trap_emul_gva_lockless_end(struct kvm_vcpu *vcpu) kvm_vcpu 1183 arch/mips/kvm/trap_emul.c struct kvm_vcpu *vcpu) kvm_vcpu 1226 arch/mips/kvm/trap_emul.c static int kvm_trap_emul_vcpu_run(struct kvm_run *run, struct kvm_vcpu *vcpu) kvm_vcpu 36 arch/mips/kvm/vz.c static struct kvm_vcpu *last_vcpu[NR_CPUS]; kvm_vcpu 38 arch/mips/kvm/vz.c static struct kvm_vcpu *last_exec_vcpu[NR_CPUS]; kvm_vcpu 81 arch/mips/kvm/vz.c static inline unsigned int kvm_vz_config_guest_wrmask(struct kvm_vcpu *vcpu) kvm_vcpu 86 arch/mips/kvm/vz.c static inline unsigned int kvm_vz_config1_guest_wrmask(struct kvm_vcpu *vcpu) kvm_vcpu 91 arch/mips/kvm/vz.c static inline unsigned int kvm_vz_config2_guest_wrmask(struct kvm_vcpu *vcpu) kvm_vcpu 96 arch/mips/kvm/vz.c static inline unsigned int kvm_vz_config3_guest_wrmask(struct kvm_vcpu *vcpu) kvm_vcpu 101 arch/mips/kvm/vz.c static inline unsigned int kvm_vz_config4_guest_wrmask(struct kvm_vcpu *vcpu) kvm_vcpu 107 arch/mips/kvm/vz.c static inline unsigned int kvm_vz_config5_guest_wrmask(struct kvm_vcpu *vcpu) kvm_vcpu 140 arch/mips/kvm/vz.c static inline unsigned int kvm_vz_config_user_wrmask(struct kvm_vcpu *vcpu) kvm_vcpu 145 arch/mips/kvm/vz.c static inline unsigned int kvm_vz_config1_user_wrmask(struct kvm_vcpu *vcpu) kvm_vcpu 156 arch/mips/kvm/vz.c static inline unsigned int kvm_vz_config2_user_wrmask(struct kvm_vcpu *vcpu) kvm_vcpu 161 arch/mips/kvm/vz.c static inline unsigned int kvm_vz_config3_user_wrmask(struct kvm_vcpu *vcpu) kvm_vcpu 173 arch/mips/kvm/vz.c static inline unsigned int kvm_vz_config4_user_wrmask(struct kvm_vcpu *vcpu) kvm_vcpu 178 arch/mips/kvm/vz.c static inline unsigned int kvm_vz_config5_user_wrmask(struct kvm_vcpu *vcpu) kvm_vcpu 189 arch/mips/kvm/vz.c static void kvm_vz_queue_irq(struct kvm_vcpu *vcpu, unsigned int priority) kvm_vcpu 195 arch/mips/kvm/vz.c static void kvm_vz_dequeue_irq(struct kvm_vcpu *vcpu, unsigned int priority) kvm_vcpu 201 arch/mips/kvm/vz.c static void kvm_vz_queue_timer_int_cb(struct kvm_vcpu *vcpu) kvm_vcpu 210 arch/mips/kvm/vz.c static void kvm_vz_dequeue_timer_int_cb(struct kvm_vcpu *vcpu) kvm_vcpu 219 arch/mips/kvm/vz.c static void kvm_vz_queue_io_int_cb(struct kvm_vcpu *vcpu, kvm_vcpu 247 arch/mips/kvm/vz.c static void kvm_vz_dequeue_io_int_cb(struct kvm_vcpu *vcpu, kvm_vcpu 282 arch/mips/kvm/vz.c static int kvm_vz_irq_deliver_cb(struct kvm_vcpu *vcpu, unsigned int priority, kvm_vcpu 310 arch/mips/kvm/vz.c static int kvm_vz_irq_clear_cb(struct kvm_vcpu *vcpu, unsigned int priority, kvm_vcpu 364 arch/mips/kvm/vz.c static bool kvm_vz_should_use_htimer(struct kvm_vcpu *vcpu) kvm_vcpu 389 arch/mips/kvm/vz.c static void _kvm_vz_restore_stimer(struct kvm_vcpu *vcpu, u32 compare, kvm_vcpu 411 arch/mips/kvm/vz.c static void _kvm_vz_restore_htimer(struct kvm_vcpu *vcpu, kvm_vcpu 448 arch/mips/kvm/vz.c static void kvm_vz_restore_timer(struct kvm_vcpu *vcpu) kvm_vcpu 469 arch/mips/kvm/vz.c void kvm_vz_acquire_htimer(struct kvm_vcpu *vcpu) kvm_vcpu 493 arch/mips/kvm/vz.c static void _kvm_vz_save_htimer(struct kvm_vcpu *vcpu, kvm_vcpu 543 arch/mips/kvm/vz.c static void kvm_vz_save_timer(struct kvm_vcpu *vcpu) kvm_vcpu 572 arch/mips/kvm/vz.c void kvm_vz_lose_htimer(struct kvm_vcpu *vcpu) kvm_vcpu 642 arch/mips/kvm/vz.c static bool is_eva_am_mapped(struct kvm_vcpu *vcpu, unsigned int am, bool eu) kvm_vcpu 714 arch/mips/kvm/vz.c static int kvm_vz_gva_to_gpa(struct kvm_vcpu *vcpu, unsigned long gva, kvm_vcpu 821 arch/mips/kvm/vz.c static int kvm_vz_badvaddr_to_gpa(struct kvm_vcpu *vcpu, unsigned long badvaddr, kvm_vcpu 842 arch/mips/kvm/vz.c static int kvm_trap_vz_no_handler(struct kvm_vcpu *vcpu) kvm_vcpu 889 arch/mips/kvm/vz.c static void kvm_write_maari(struct kvm_vcpu *vcpu, unsigned long val) kvm_vcpu 903 arch/mips/kvm/vz.c struct kvm_vcpu *vcpu) kvm_vcpu 1066 arch/mips/kvm/vz.c struct kvm_vcpu *vcpu) kvm_vcpu 1133 arch/mips/kvm/vz.c struct kvm_vcpu *vcpu) kvm_vcpu 1211 arch/mips/kvm/vz.c struct kvm_vcpu *vcpu) kvm_vcpu 1337 arch/mips/kvm/vz.c struct kvm_vcpu *vcpu) kvm_vcpu 1349 arch/mips/kvm/vz.c struct kvm_vcpu *vcpu) kvm_vcpu 1381 arch/mips/kvm/vz.c struct kvm_vcpu *vcpu) kvm_vcpu 1398 arch/mips/kvm/vz.c static int kvm_trap_vz_handle_guest_exit(struct kvm_vcpu *vcpu) kvm_vcpu 1466 arch/mips/kvm/vz.c static int kvm_trap_vz_handle_cop_unusable(struct kvm_vcpu *vcpu) kvm_vcpu 1513 arch/mips/kvm/vz.c static int kvm_trap_vz_handle_msa_disabled(struct kvm_vcpu *vcpu) kvm_vcpu 1536 arch/mips/kvm/vz.c static int kvm_trap_vz_handle_tlb_ld_miss(struct kvm_vcpu *vcpu) kvm_vcpu 1583 arch/mips/kvm/vz.c static int kvm_trap_vz_handle_tlb_st_miss(struct kvm_vcpu *vcpu) kvm_vcpu 1694 arch/mips/kvm/vz.c static unsigned long kvm_vz_num_regs(struct kvm_vcpu *vcpu) kvm_vcpu 1718 arch/mips/kvm/vz.c static int kvm_vz_copy_reg_indices(struct kvm_vcpu *vcpu, u64 __user *indices) kvm_vcpu 1822 arch/mips/kvm/vz.c static int kvm_vz_get_one_reg(struct kvm_vcpu *vcpu, kvm_vcpu 2043 arch/mips/kvm/vz.c static int kvm_vz_set_one_reg(struct kvm_vcpu *vcpu, kvm_vcpu 2313 arch/mips/kvm/vz.c static void kvm_vz_get_new_guestid(unsigned long cpu, struct kvm_vcpu *vcpu) kvm_vcpu 2335 arch/mips/kvm/vz.c static int kvm_vz_check_requests(struct kvm_vcpu *vcpu, int cpu) kvm_vcpu 2362 arch/mips/kvm/vz.c static void kvm_vz_vcpu_save_wired(struct kvm_vcpu *vcpu) kvm_vcpu 2395 arch/mips/kvm/vz.c static void kvm_vz_vcpu_load_wired(struct kvm_vcpu *vcpu) kvm_vcpu 2403 arch/mips/kvm/vz.c static void kvm_vz_vcpu_load_tlb(struct kvm_vcpu *vcpu, int cpu) kvm_vcpu 2464 arch/mips/kvm/vz.c static int kvm_vz_vcpu_load(struct kvm_vcpu *vcpu, int cpu) kvm_vcpu 2606 arch/mips/kvm/vz.c static int kvm_vz_vcpu_put(struct kvm_vcpu *vcpu, int cpu) kvm_vcpu 2938 arch/mips/kvm/vz.c static int kvm_vz_vcpu_init(struct kvm_vcpu *vcpu) kvm_vcpu 2948 arch/mips/kvm/vz.c static void kvm_vz_vcpu_uninit(struct kvm_vcpu *vcpu) kvm_vcpu 2965 arch/mips/kvm/vz.c static int kvm_vz_vcpu_setup(struct kvm_vcpu *vcpu) kvm_vcpu 3132 arch/mips/kvm/vz.c static void kvm_vz_vcpu_reenter(struct kvm_run *run, struct kvm_vcpu *vcpu) kvm_vcpu 3148 arch/mips/kvm/vz.c static int kvm_vz_vcpu_run(struct kvm_run *run, struct kvm_vcpu *vcpu) kvm_vcpu 148 arch/powerpc/include/asm/asm-prototypes.h struct kvm_vcpu; kvm_vcpu 149 arch/powerpc/include/asm/asm-prototypes.h void _kvmppc_restore_tm_pr(struct kvm_vcpu *vcpu, u64 guest_msr); kvm_vcpu 150 arch/powerpc/include/asm/asm-prototypes.h void _kvmppc_save_tm_pr(struct kvm_vcpu *vcpu, u64 guest_msr); kvm_vcpu 163 arch/powerpc/include/asm/asm-prototypes.h void kvmppc_save_tm_hv(struct kvm_vcpu *vcpu, u64 msr, bool preserve_nv); kvm_vcpu 164 arch/powerpc/include/asm/asm-prototypes.h void kvmppc_restore_tm_hv(struct kvm_vcpu *vcpu, u64 msr, bool preserve_nv); kvm_vcpu 166 arch/powerpc/include/asm/asm-prototypes.h static inline void kvmppc_save_tm_hv(struct kvm_vcpu *vcpu, u64 msr, kvm_vcpu 168 arch/powerpc/include/asm/asm-prototypes.h static inline void kvmppc_restore_tm_hv(struct kvm_vcpu *vcpu, u64 msr, kvm_vcpu 174 arch/powerpc/include/asm/asm-prototypes.h void kvmhv_save_guest_pmu(struct kvm_vcpu *vcpu, bool pmu_in_use); kvm_vcpu 175 arch/powerpc/include/asm/asm-prototypes.h void kvmhv_load_guest_pmu(struct kvm_vcpu *vcpu); kvm_vcpu 177 arch/powerpc/include/asm/asm-prototypes.h int __kvmhv_vcpu_entry_p9(struct kvm_vcpu *vcpu); kvm_vcpu 179 arch/powerpc/include/asm/asm-prototypes.h long kvmppc_h_set_dabr(struct kvm_vcpu *vcpu, unsigned long dabr); kvm_vcpu 180 arch/powerpc/include/asm/asm-prototypes.h long kvmppc_h_set_xdabr(struct kvm_vcpu *vcpu, unsigned long dabr, kvm_vcpu 78 arch/powerpc/include/asm/kvm_book3s.h struct kvm_vcpu *runnable_threads[MAX_SMT_THREADS]; kvm_vcpu 85 arch/powerpc/include/asm/kvm_book3s.h struct kvm_vcpu *runner; kvm_vcpu 145 arch/powerpc/include/asm/kvm_book3s.h extern void kvmppc_mmu_pte_flush(struct kvm_vcpu *vcpu, ulong ea, ulong ea_mask); kvm_vcpu 146 arch/powerpc/include/asm/kvm_book3s.h extern void kvmppc_mmu_pte_vflush(struct kvm_vcpu *vcpu, u64 vp, u64 vp_mask); kvm_vcpu 147 arch/powerpc/include/asm/kvm_book3s.h extern void kvmppc_mmu_pte_pflush(struct kvm_vcpu *vcpu, ulong pa_start, ulong pa_end); kvm_vcpu 148 arch/powerpc/include/asm/kvm_book3s.h extern void kvmppc_set_msr(struct kvm_vcpu *vcpu, u64 new_msr); kvm_vcpu 149 arch/powerpc/include/asm/kvm_book3s.h extern void kvmppc_mmu_book3s_64_init(struct kvm_vcpu *vcpu); kvm_vcpu 150 arch/powerpc/include/asm/kvm_book3s.h extern void kvmppc_mmu_book3s_32_init(struct kvm_vcpu *vcpu); kvm_vcpu 151 arch/powerpc/include/asm/kvm_book3s.h extern void kvmppc_mmu_book3s_hv_init(struct kvm_vcpu *vcpu); kvm_vcpu 152 arch/powerpc/include/asm/kvm_book3s.h extern int kvmppc_mmu_map_page(struct kvm_vcpu *vcpu, struct kvmppc_pte *pte, kvm_vcpu 154 arch/powerpc/include/asm/kvm_book3s.h extern void kvmppc_mmu_unmap_page(struct kvm_vcpu *vcpu, struct kvmppc_pte *pte); kvm_vcpu 155 arch/powerpc/include/asm/kvm_book3s.h extern int kvmppc_mmu_map_segment(struct kvm_vcpu *vcpu, ulong eaddr); kvm_vcpu 156 arch/powerpc/include/asm/kvm_book3s.h extern void kvmppc_mmu_flush_segment(struct kvm_vcpu *vcpu, ulong eaddr, ulong seg_size); kvm_vcpu 157 arch/powerpc/include/asm/kvm_book3s.h extern void kvmppc_mmu_flush_segments(struct kvm_vcpu *vcpu); kvm_vcpu 159 arch/powerpc/include/asm/kvm_book3s.h struct kvm_vcpu *vcpu, unsigned long addr, kvm_vcpu 163 arch/powerpc/include/asm/kvm_book3s.h extern int kvmppc_hv_emulate_mmio(struct kvm_run *run, struct kvm_vcpu *vcpu, kvm_vcpu 166 arch/powerpc/include/asm/kvm_book3s.h extern void kvmppc_mmu_hpte_cache_map(struct kvm_vcpu *vcpu, struct hpte_cache *pte); kvm_vcpu 167 arch/powerpc/include/asm/kvm_book3s.h extern struct hpte_cache *kvmppc_mmu_hpte_cache_next(struct kvm_vcpu *vcpu); kvm_vcpu 169 arch/powerpc/include/asm/kvm_book3s.h extern void kvmppc_mmu_hpte_destroy(struct kvm_vcpu *vcpu); kvm_vcpu 170 arch/powerpc/include/asm/kvm_book3s.h extern int kvmppc_mmu_hpte_init(struct kvm_vcpu *vcpu); kvm_vcpu 171 arch/powerpc/include/asm/kvm_book3s.h extern void kvmppc_mmu_invalidate_pte(struct kvm_vcpu *vcpu, struct hpte_cache *pte); kvm_vcpu 178 arch/powerpc/include/asm/kvm_book3s.h struct kvm_vcpu *vcpu, kvm_vcpu 183 arch/powerpc/include/asm/kvm_book3s.h extern long kvmhv_copy_from_guest_radix(struct kvm_vcpu *vcpu, gva_t eaddr, kvm_vcpu 185 arch/powerpc/include/asm/kvm_book3s.h extern long kvmhv_copy_to_guest_radix(struct kvm_vcpu *vcpu, gva_t eaddr, kvm_vcpu 187 arch/powerpc/include/asm/kvm_book3s.h extern int kvmppc_mmu_walk_radix_tree(struct kvm_vcpu *vcpu, gva_t eaddr, kvm_vcpu 190 arch/powerpc/include/asm/kvm_book3s.h extern int kvmppc_mmu_radix_translate_table(struct kvm_vcpu *vcpu, gva_t eaddr, kvm_vcpu 193 arch/powerpc/include/asm/kvm_book3s.h extern int kvmppc_mmu_radix_xlate(struct kvm_vcpu *vcpu, gva_t eaddr, kvm_vcpu 204 arch/powerpc/include/asm/kvm_book3s.h extern int kvmppc_book3s_instantiate_page(struct kvm_vcpu *vcpu, kvm_vcpu 228 arch/powerpc/include/asm/kvm_book3s.h extern int kvmppc_ld(struct kvm_vcpu *vcpu, ulong *eaddr, int size, void *ptr, bool data); kvm_vcpu 229 arch/powerpc/include/asm/kvm_book3s.h extern void kvmppc_book3s_queue_irqprio(struct kvm_vcpu *vcpu, unsigned int vec); kvm_vcpu 230 arch/powerpc/include/asm/kvm_book3s.h extern void kvmppc_book3s_dequeue_irqprio(struct kvm_vcpu *vcpu, kvm_vcpu 232 arch/powerpc/include/asm/kvm_book3s.h extern void kvmppc_inject_interrupt(struct kvm_vcpu *vcpu, int vec, u64 flags); kvm_vcpu 233 arch/powerpc/include/asm/kvm_book3s.h extern void kvmppc_trigger_fac_interrupt(struct kvm_vcpu *vcpu, ulong fac); kvm_vcpu 234 arch/powerpc/include/asm/kvm_book3s.h extern void kvmppc_set_bat(struct kvm_vcpu *vcpu, struct kvmppc_bat *bat, kvm_vcpu 236 arch/powerpc/include/asm/kvm_book3s.h extern void kvmppc_giveup_ext(struct kvm_vcpu *vcpu, ulong msr); kvm_vcpu 237 arch/powerpc/include/asm/kvm_book3s.h extern int kvmppc_emulate_paired_single(struct kvm_run *run, struct kvm_vcpu *vcpu); kvm_vcpu 238 arch/powerpc/include/asm/kvm_book3s.h extern kvm_pfn_t kvmppc_gpa_to_pfn(struct kvm_vcpu *vcpu, gpa_t gpa, kvm_vcpu 265 arch/powerpc/include/asm/kvm_book3s.h extern void kvmppc_set_fscr(struct kvm_vcpu *vcpu, u64 fscr); kvm_vcpu 267 arch/powerpc/include/asm/kvm_book3s.h extern int kvmhv_p9_tm_emulation_early(struct kvm_vcpu *vcpu); kvm_vcpu 268 arch/powerpc/include/asm/kvm_book3s.h extern int kvmhv_p9_tm_emulation(struct kvm_vcpu *vcpu); kvm_vcpu 269 arch/powerpc/include/asm/kvm_book3s.h extern void kvmhv_emulate_tm_rollback(struct kvm_vcpu *vcpu); kvm_vcpu 273 arch/powerpc/include/asm/kvm_book3s.h extern u32 kvmppc_alignment_dsisr(struct kvm_vcpu *vcpu, unsigned int inst); kvm_vcpu 274 arch/powerpc/include/asm/kvm_book3s.h extern ulong kvmppc_alignment_dar(struct kvm_vcpu *vcpu, unsigned int inst); kvm_vcpu 275 arch/powerpc/include/asm/kvm_book3s.h extern int kvmppc_h_pr(struct kvm_vcpu *vcpu, unsigned long cmd); kvm_vcpu 279 arch/powerpc/include/asm/kvm_book3s.h extern void kvmppc_copy_to_svcpu(struct kvm_vcpu *vcpu); kvm_vcpu 280 arch/powerpc/include/asm/kvm_book3s.h extern void kvmppc_copy_from_svcpu(struct kvm_vcpu *vcpu); kvm_vcpu 283 arch/powerpc/include/asm/kvm_book3s.h void kvmppc_save_tm_pr(struct kvm_vcpu *vcpu); kvm_vcpu 284 arch/powerpc/include/asm/kvm_book3s.h void kvmppc_restore_tm_pr(struct kvm_vcpu *vcpu); kvm_vcpu 285 arch/powerpc/include/asm/kvm_book3s.h void kvmppc_save_tm_sprs(struct kvm_vcpu *vcpu); kvm_vcpu 286 arch/powerpc/include/asm/kvm_book3s.h void kvmppc_restore_tm_sprs(struct kvm_vcpu *vcpu); kvm_vcpu 288 arch/powerpc/include/asm/kvm_book3s.h static inline void kvmppc_save_tm_pr(struct kvm_vcpu *vcpu) {} kvm_vcpu 289 arch/powerpc/include/asm/kvm_book3s.h static inline void kvmppc_restore_tm_pr(struct kvm_vcpu *vcpu) {} kvm_vcpu 290 arch/powerpc/include/asm/kvm_book3s.h static inline void kvmppc_save_tm_sprs(struct kvm_vcpu *vcpu) {} kvm_vcpu 291 arch/powerpc/include/asm/kvm_book3s.h static inline void kvmppc_restore_tm_sprs(struct kvm_vcpu *vcpu) {} kvm_vcpu 297 arch/powerpc/include/asm/kvm_book3s.h long kvmhv_set_partition_table(struct kvm_vcpu *vcpu); kvm_vcpu 298 arch/powerpc/include/asm/kvm_book3s.h long kvmhv_copy_tofrom_guest_nested(struct kvm_vcpu *vcpu); kvm_vcpu 301 arch/powerpc/include/asm/kvm_book3s.h long kvmhv_enter_nested_guest(struct kvm_vcpu *vcpu); kvm_vcpu 302 arch/powerpc/include/asm/kvm_book3s.h long kvmhv_do_nested_tlbie(struct kvm_vcpu *vcpu); kvm_vcpu 303 arch/powerpc/include/asm/kvm_book3s.h int kvmhv_run_single_vcpu(struct kvm_run *kvm_run, struct kvm_vcpu *vcpu, kvm_vcpu 305 arch/powerpc/include/asm/kvm_book3s.h void kvmhv_save_hv_regs(struct kvm_vcpu *vcpu, struct hv_guest_state *hr); kvm_vcpu 306 arch/powerpc/include/asm/kvm_book3s.h void kvmhv_restore_hv_return_state(struct kvm_vcpu *vcpu, kvm_vcpu 308 arch/powerpc/include/asm/kvm_book3s.h long int kvmhv_nested_page_fault(struct kvm_run *run, struct kvm_vcpu *vcpu); kvm_vcpu 310 arch/powerpc/include/asm/kvm_book3s.h void kvmppc_giveup_fac(struct kvm_vcpu *vcpu, ulong fac); kvm_vcpu 314 arch/powerpc/include/asm/kvm_book3s.h static inline struct kvmppc_vcpu_book3s *to_book3s(struct kvm_vcpu *vcpu) kvm_vcpu 328 arch/powerpc/include/asm/kvm_book3s.h static inline void kvmppc_set_gpr(struct kvm_vcpu *vcpu, int num, ulong val) kvm_vcpu 333 arch/powerpc/include/asm/kvm_book3s.h static inline ulong kvmppc_get_gpr(struct kvm_vcpu *vcpu, int num) kvm_vcpu 338 arch/powerpc/include/asm/kvm_book3s.h static inline void kvmppc_set_cr(struct kvm_vcpu *vcpu, u32 val) kvm_vcpu 343 arch/powerpc/include/asm/kvm_book3s.h static inline u32 kvmppc_get_cr(struct kvm_vcpu *vcpu) kvm_vcpu 348 arch/powerpc/include/asm/kvm_book3s.h static inline void kvmppc_set_xer(struct kvm_vcpu *vcpu, ulong val) kvm_vcpu 353 arch/powerpc/include/asm/kvm_book3s.h static inline ulong kvmppc_get_xer(struct kvm_vcpu *vcpu) kvm_vcpu 358 arch/powerpc/include/asm/kvm_book3s.h static inline void kvmppc_set_ctr(struct kvm_vcpu *vcpu, ulong val) kvm_vcpu 363 arch/powerpc/include/asm/kvm_book3s.h static inline ulong kvmppc_get_ctr(struct kvm_vcpu *vcpu) kvm_vcpu 368 arch/powerpc/include/asm/kvm_book3s.h static inline void kvmppc_set_lr(struct kvm_vcpu *vcpu, ulong val) kvm_vcpu 373 arch/powerpc/include/asm/kvm_book3s.h static inline ulong kvmppc_get_lr(struct kvm_vcpu *vcpu) kvm_vcpu 378 arch/powerpc/include/asm/kvm_book3s.h static inline void kvmppc_set_pc(struct kvm_vcpu *vcpu, ulong val) kvm_vcpu 383 arch/powerpc/include/asm/kvm_book3s.h static inline ulong kvmppc_get_pc(struct kvm_vcpu *vcpu) kvm_vcpu 388 arch/powerpc/include/asm/kvm_book3s.h static inline u64 kvmppc_get_msr(struct kvm_vcpu *vcpu); kvm_vcpu 389 arch/powerpc/include/asm/kvm_book3s.h static inline bool kvmppc_need_byteswap(struct kvm_vcpu *vcpu) kvm_vcpu 394 arch/powerpc/include/asm/kvm_book3s.h static inline ulong kvmppc_get_fault_dar(struct kvm_vcpu *vcpu) kvm_vcpu 405 arch/powerpc/include/asm/kvm_book3s.h static inline bool kvmppc_supports_magic_page(struct kvm_vcpu *vcpu) kvm_vcpu 411 arch/powerpc/include/asm/kvm_book3s.h extern int kvmppc_h_logical_ci_load(struct kvm_vcpu *vcpu); kvm_vcpu 412 arch/powerpc/include/asm/kvm_book3s.h extern int kvmppc_h_logical_ci_store(struct kvm_vcpu *vcpu); kvm_vcpu 12 arch/powerpc/include/asm/kvm_book3s_32.h static inline struct kvmppc_book3s_shadow_vcpu *svcpu_get(struct kvm_vcpu *vcpu) kvm_vcpu 124 arch/powerpc/include/asm/kvm_book3s_64.h static inline struct kvmppc_book3s_shadow_vcpu *svcpu_get(struct kvm_vcpu *vcpu) kvm_vcpu 143 arch/powerpc/include/asm/kvm_book3s_64.h static inline bool kvmhv_vcpu_is_radix(struct kvm_vcpu *vcpu) kvm_vcpu 588 arch/powerpc/include/asm/kvm_book3s_64.h static inline void copy_from_checkpoint(struct kvm_vcpu *vcpu) kvm_vcpu 605 arch/powerpc/include/asm/kvm_book3s_64.h static inline void copy_to_checkpoint(struct kvm_vcpu *vcpu) kvm_vcpu 115 arch/powerpc/include/asm/kvm_book3s_asm.h struct kvm_vcpu *kvm_vcpu; kvm_vcpu 26 arch/powerpc/include/asm/kvm_booke.h static inline void kvmppc_set_gpr(struct kvm_vcpu *vcpu, int num, ulong val) kvm_vcpu 31 arch/powerpc/include/asm/kvm_booke.h static inline ulong kvmppc_get_gpr(struct kvm_vcpu *vcpu, int num) kvm_vcpu 36 arch/powerpc/include/asm/kvm_booke.h static inline void kvmppc_set_cr(struct kvm_vcpu *vcpu, u32 val) kvm_vcpu 41 arch/powerpc/include/asm/kvm_booke.h static inline u32 kvmppc_get_cr(struct kvm_vcpu *vcpu) kvm_vcpu 46 arch/powerpc/include/asm/kvm_booke.h static inline void kvmppc_set_xer(struct kvm_vcpu *vcpu, ulong val) kvm_vcpu 51 arch/powerpc/include/asm/kvm_booke.h static inline ulong kvmppc_get_xer(struct kvm_vcpu *vcpu) kvm_vcpu 56 arch/powerpc/include/asm/kvm_booke.h static inline bool kvmppc_need_byteswap(struct kvm_vcpu *vcpu) kvm_vcpu 62 arch/powerpc/include/asm/kvm_booke.h static inline void kvmppc_set_ctr(struct kvm_vcpu *vcpu, ulong val) kvm_vcpu 67 arch/powerpc/include/asm/kvm_booke.h static inline ulong kvmppc_get_ctr(struct kvm_vcpu *vcpu) kvm_vcpu 72 arch/powerpc/include/asm/kvm_booke.h static inline void kvmppc_set_lr(struct kvm_vcpu *vcpu, ulong val) kvm_vcpu 77 arch/powerpc/include/asm/kvm_booke.h static inline ulong kvmppc_get_lr(struct kvm_vcpu *vcpu) kvm_vcpu 82 arch/powerpc/include/asm/kvm_booke.h static inline void kvmppc_set_pc(struct kvm_vcpu *vcpu, ulong val) kvm_vcpu 87 arch/powerpc/include/asm/kvm_booke.h static inline ulong kvmppc_get_pc(struct kvm_vcpu *vcpu) kvm_vcpu 92 arch/powerpc/include/asm/kvm_booke.h static inline ulong kvmppc_get_fault_dar(struct kvm_vcpu *vcpu) kvm_vcpu 97 arch/powerpc/include/asm/kvm_booke.h static inline bool kvmppc_supports_magic_page(struct kvm_vcpu *vcpu) kvm_vcpu 393 arch/powerpc/include/asm/kvm_host.h void (*slbmte)(struct kvm_vcpu *vcpu, u64 rb, u64 rs); kvm_vcpu 394 arch/powerpc/include/asm/kvm_host.h u64 (*slbmfee)(struct kvm_vcpu *vcpu, u64 slb_nr); kvm_vcpu 395 arch/powerpc/include/asm/kvm_host.h u64 (*slbmfev)(struct kvm_vcpu *vcpu, u64 slb_nr); kvm_vcpu 396 arch/powerpc/include/asm/kvm_host.h int (*slbfee)(struct kvm_vcpu *vcpu, gva_t eaddr, ulong *ret_slb); kvm_vcpu 397 arch/powerpc/include/asm/kvm_host.h void (*slbie)(struct kvm_vcpu *vcpu, u64 slb_nr); kvm_vcpu 398 arch/powerpc/include/asm/kvm_host.h void (*slbia)(struct kvm_vcpu *vcpu); kvm_vcpu 400 arch/powerpc/include/asm/kvm_host.h void (*mtsrin)(struct kvm_vcpu *vcpu, u32 srnum, ulong value); kvm_vcpu 401 arch/powerpc/include/asm/kvm_host.h u32 (*mfsrin)(struct kvm_vcpu *vcpu, u32 srnum); kvm_vcpu 402 arch/powerpc/include/asm/kvm_host.h int (*xlate)(struct kvm_vcpu *vcpu, gva_t eaddr, kvm_vcpu 404 arch/powerpc/include/asm/kvm_host.h void (*reset_msr)(struct kvm_vcpu *vcpu); kvm_vcpu 405 arch/powerpc/include/asm/kvm_host.h void (*tlbie)(struct kvm_vcpu *vcpu, ulong addr, bool large); kvm_vcpu 406 arch/powerpc/include/asm/kvm_host.h int (*esid_to_vsid)(struct kvm_vcpu *vcpu, ulong esid, u64 *vsid); kvm_vcpu 407 arch/powerpc/include/asm/kvm_host.h u64 (*ea_to_vp)(struct kvm_vcpu *vcpu, gva_t eaddr, bool data); kvm_vcpu 408 arch/powerpc/include/asm/kvm_host.h bool (*is_dcbz32)(struct kvm_vcpu *vcpu); kvm_vcpu 860 arch/powerpc/include/asm/kvm_host.h static inline void kvm_arch_sched_in(struct kvm_vcpu *vcpu, int cpu) {} kvm_vcpu 862 arch/powerpc/include/asm/kvm_host.h static inline void kvm_arch_vcpu_blocking(struct kvm_vcpu *vcpu) {} kvm_vcpu 863 arch/powerpc/include/asm/kvm_host.h static inline void kvm_arch_vcpu_unblocking(struct kvm_vcpu *vcpu) {} kvm_vcpu 864 arch/powerpc/include/asm/kvm_host.h static inline void kvm_arch_vcpu_block_finish(struct kvm_vcpu *vcpu) {} kvm_vcpu 61 arch/powerpc/include/asm/kvm_ppc.h extern int kvmppc_vcpu_run(struct kvm_run *kvm_run, struct kvm_vcpu *vcpu); kvm_vcpu 62 arch/powerpc/include/asm/kvm_ppc.h extern int __kvmppc_vcpu_run(struct kvm_run *kvm_run, struct kvm_vcpu *vcpu); kvm_vcpu 65 arch/powerpc/include/asm/kvm_ppc.h extern void kvmppc_dump_vcpu(struct kvm_vcpu *vcpu); kvm_vcpu 66 arch/powerpc/include/asm/kvm_ppc.h extern int kvmppc_handle_load(struct kvm_run *run, struct kvm_vcpu *vcpu, kvm_vcpu 69 arch/powerpc/include/asm/kvm_ppc.h extern int kvmppc_handle_loads(struct kvm_run *run, struct kvm_vcpu *vcpu, kvm_vcpu 72 arch/powerpc/include/asm/kvm_ppc.h extern int kvmppc_handle_vsx_load(struct kvm_run *run, struct kvm_vcpu *vcpu, kvm_vcpu 75 arch/powerpc/include/asm/kvm_ppc.h extern int kvmppc_handle_vmx_load(struct kvm_run *run, struct kvm_vcpu *vcpu, kvm_vcpu 77 arch/powerpc/include/asm/kvm_ppc.h extern int kvmppc_handle_vmx_store(struct kvm_run *run, struct kvm_vcpu *vcpu, kvm_vcpu 79 arch/powerpc/include/asm/kvm_ppc.h extern int kvmppc_handle_store(struct kvm_run *run, struct kvm_vcpu *vcpu, kvm_vcpu 82 arch/powerpc/include/asm/kvm_ppc.h extern int kvmppc_handle_vsx_store(struct kvm_run *run, struct kvm_vcpu *vcpu, kvm_vcpu 86 arch/powerpc/include/asm/kvm_ppc.h extern int kvmppc_load_last_inst(struct kvm_vcpu *vcpu, kvm_vcpu 89 arch/powerpc/include/asm/kvm_ppc.h extern int kvmppc_ld(struct kvm_vcpu *vcpu, ulong *eaddr, int size, void *ptr, kvm_vcpu 91 arch/powerpc/include/asm/kvm_ppc.h extern int kvmppc_st(struct kvm_vcpu *vcpu, ulong *eaddr, int size, void *ptr, kvm_vcpu 94 arch/powerpc/include/asm/kvm_ppc.h struct kvm_vcpu *vcpu); kvm_vcpu 95 arch/powerpc/include/asm/kvm_ppc.h extern int kvmppc_emulate_loadstore(struct kvm_vcpu *vcpu); kvm_vcpu 96 arch/powerpc/include/asm/kvm_ppc.h extern int kvmppc_emulate_mmio(struct kvm_run *run, struct kvm_vcpu *vcpu); kvm_vcpu 97 arch/powerpc/include/asm/kvm_ppc.h extern void kvmppc_emulate_dec(struct kvm_vcpu *vcpu); kvm_vcpu 98 arch/powerpc/include/asm/kvm_ppc.h extern u32 kvmppc_get_dec(struct kvm_vcpu *vcpu, u64 tb); kvm_vcpu 99 arch/powerpc/include/asm/kvm_ppc.h extern void kvmppc_decrementer_func(struct kvm_vcpu *vcpu); kvm_vcpu 100 arch/powerpc/include/asm/kvm_ppc.h extern int kvmppc_sanity_check(struct kvm_vcpu *vcpu); kvm_vcpu 101 arch/powerpc/include/asm/kvm_ppc.h extern int kvmppc_subarch_vcpu_init(struct kvm_vcpu *vcpu); kvm_vcpu 102 arch/powerpc/include/asm/kvm_ppc.h extern void kvmppc_subarch_vcpu_uninit(struct kvm_vcpu *vcpu); kvm_vcpu 106 arch/powerpc/include/asm/kvm_ppc.h extern void kvmppc_mmu_map(struct kvm_vcpu *vcpu, u64 gvaddr, gpa_t gpaddr, kvm_vcpu 108 arch/powerpc/include/asm/kvm_ppc.h extern void kvmppc_mmu_priv_switch(struct kvm_vcpu *vcpu, int usermode); kvm_vcpu 109 arch/powerpc/include/asm/kvm_ppc.h extern void kvmppc_mmu_switch_pid(struct kvm_vcpu *vcpu, u32 pid); kvm_vcpu 110 arch/powerpc/include/asm/kvm_ppc.h extern void kvmppc_mmu_destroy(struct kvm_vcpu *vcpu); kvm_vcpu 111 arch/powerpc/include/asm/kvm_ppc.h extern int kvmppc_mmu_init(struct kvm_vcpu *vcpu); kvm_vcpu 112 arch/powerpc/include/asm/kvm_ppc.h extern int kvmppc_mmu_dtlb_index(struct kvm_vcpu *vcpu, gva_t eaddr); kvm_vcpu 113 arch/powerpc/include/asm/kvm_ppc.h extern int kvmppc_mmu_itlb_index(struct kvm_vcpu *vcpu, gva_t eaddr); kvm_vcpu 114 arch/powerpc/include/asm/kvm_ppc.h extern gpa_t kvmppc_mmu_xlate(struct kvm_vcpu *vcpu, unsigned int gtlb_index, kvm_vcpu 116 arch/powerpc/include/asm/kvm_ppc.h extern void kvmppc_mmu_dtlb_miss(struct kvm_vcpu *vcpu); kvm_vcpu 117 arch/powerpc/include/asm/kvm_ppc.h extern void kvmppc_mmu_itlb_miss(struct kvm_vcpu *vcpu); kvm_vcpu 118 arch/powerpc/include/asm/kvm_ppc.h extern int kvmppc_xlate(struct kvm_vcpu *vcpu, ulong eaddr, kvm_vcpu 122 arch/powerpc/include/asm/kvm_ppc.h extern struct kvm_vcpu *kvmppc_core_vcpu_create(struct kvm *kvm, kvm_vcpu 124 arch/powerpc/include/asm/kvm_ppc.h extern void kvmppc_core_vcpu_free(struct kvm_vcpu *vcpu); kvm_vcpu 125 arch/powerpc/include/asm/kvm_ppc.h extern int kvmppc_core_vcpu_setup(struct kvm_vcpu *vcpu); kvm_vcpu 127 arch/powerpc/include/asm/kvm_ppc.h extern int kvmppc_core_vcpu_translate(struct kvm_vcpu *vcpu, kvm_vcpu 130 arch/powerpc/include/asm/kvm_ppc.h extern void kvmppc_core_vcpu_load(struct kvm_vcpu *vcpu, int cpu); kvm_vcpu 131 arch/powerpc/include/asm/kvm_ppc.h extern void kvmppc_core_vcpu_put(struct kvm_vcpu *vcpu); kvm_vcpu 133 arch/powerpc/include/asm/kvm_ppc.h extern int kvmppc_core_prepare_to_enter(struct kvm_vcpu *vcpu); kvm_vcpu 134 arch/powerpc/include/asm/kvm_ppc.h extern int kvmppc_core_pending_dec(struct kvm_vcpu *vcpu); kvm_vcpu 135 arch/powerpc/include/asm/kvm_ppc.h extern void kvmppc_core_queue_machine_check(struct kvm_vcpu *vcpu, ulong flags); kvm_vcpu 136 arch/powerpc/include/asm/kvm_ppc.h extern void kvmppc_core_queue_program(struct kvm_vcpu *vcpu, ulong flags); kvm_vcpu 137 arch/powerpc/include/asm/kvm_ppc.h extern void kvmppc_core_queue_fpunavail(struct kvm_vcpu *vcpu); kvm_vcpu 138 arch/powerpc/include/asm/kvm_ppc.h extern void kvmppc_core_queue_vec_unavail(struct kvm_vcpu *vcpu); kvm_vcpu 139 arch/powerpc/include/asm/kvm_ppc.h extern void kvmppc_core_queue_vsx_unavail(struct kvm_vcpu *vcpu); kvm_vcpu 140 arch/powerpc/include/asm/kvm_ppc.h extern void kvmppc_core_queue_dec(struct kvm_vcpu *vcpu); kvm_vcpu 141 arch/powerpc/include/asm/kvm_ppc.h extern void kvmppc_core_dequeue_dec(struct kvm_vcpu *vcpu); kvm_vcpu 142 arch/powerpc/include/asm/kvm_ppc.h extern void kvmppc_core_queue_external(struct kvm_vcpu *vcpu, kvm_vcpu 144 arch/powerpc/include/asm/kvm_ppc.h extern void kvmppc_core_dequeue_external(struct kvm_vcpu *vcpu); kvm_vcpu 145 arch/powerpc/include/asm/kvm_ppc.h extern void kvmppc_core_queue_dtlb_miss(struct kvm_vcpu *vcpu, ulong dear_flags, kvm_vcpu 147 arch/powerpc/include/asm/kvm_ppc.h extern void kvmppc_core_queue_data_storage(struct kvm_vcpu *vcpu, kvm_vcpu 150 arch/powerpc/include/asm/kvm_ppc.h extern void kvmppc_core_queue_itlb_miss(struct kvm_vcpu *vcpu); kvm_vcpu 151 arch/powerpc/include/asm/kvm_ppc.h extern void kvmppc_core_queue_inst_storage(struct kvm_vcpu *vcpu, kvm_vcpu 153 arch/powerpc/include/asm/kvm_ppc.h extern void kvmppc_core_flush_tlb(struct kvm_vcpu *vcpu); kvm_vcpu 154 arch/powerpc/include/asm/kvm_ppc.h extern int kvmppc_core_check_requests(struct kvm_vcpu *vcpu); kvm_vcpu 159 arch/powerpc/include/asm/kvm_ppc.h extern void kvmppc_core_destroy_mmu(struct kvm_vcpu *vcpu); kvm_vcpu 160 arch/powerpc/include/asm/kvm_ppc.h extern int kvmppc_kvm_pv(struct kvm_vcpu *vcpu); kvm_vcpu 161 arch/powerpc/include/asm/kvm_ppc.h extern void kvmppc_map_magic(struct kvm_vcpu *vcpu); kvm_vcpu 170 arch/powerpc/include/asm/kvm_ppc.h extern void kvmppc_map_vrma(struct kvm_vcpu *vcpu, kvm_vcpu 172 arch/powerpc/include/asm/kvm_ppc.h extern int kvmppc_pseries_do_hcall(struct kvm_vcpu *vcpu); kvm_vcpu 189 arch/powerpc/include/asm/kvm_ppc.h extern long kvmppc_h_put_tce(struct kvm_vcpu *vcpu, unsigned long liobn, kvm_vcpu 191 arch/powerpc/include/asm/kvm_ppc.h extern long kvmppc_h_put_tce_indirect(struct kvm_vcpu *vcpu, kvm_vcpu 194 arch/powerpc/include/asm/kvm_ppc.h extern long kvmppc_h_stuff_tce(struct kvm_vcpu *vcpu, kvm_vcpu 197 arch/powerpc/include/asm/kvm_ppc.h extern long kvmppc_h_get_tce(struct kvm_vcpu *vcpu, unsigned long liobn, kvm_vcpu 225 arch/powerpc/include/asm/kvm_ppc.h extern int kvmppc_prepare_to_enter(struct kvm_vcpu *vcpu); kvm_vcpu 233 arch/powerpc/include/asm/kvm_ppc.h int kvm_vcpu_ioctl_interrupt(struct kvm_vcpu *vcpu, struct kvm_interrupt *irq); kvm_vcpu 236 arch/powerpc/include/asm/kvm_ppc.h extern int kvmppc_rtas_hcall(struct kvm_vcpu *vcpu); kvm_vcpu 246 arch/powerpc/include/asm/kvm_ppc.h void kvmppc_core_dequeue_debug(struct kvm_vcpu *vcpu); kvm_vcpu 247 arch/powerpc/include/asm/kvm_ppc.h void kvmppc_core_queue_debug(struct kvm_vcpu *vcpu); kvm_vcpu 266 arch/powerpc/include/asm/kvm_ppc.h int (*get_sregs)(struct kvm_vcpu *vcpu, struct kvm_sregs *sregs); kvm_vcpu 267 arch/powerpc/include/asm/kvm_ppc.h int (*set_sregs)(struct kvm_vcpu *vcpu, struct kvm_sregs *sregs); kvm_vcpu 268 arch/powerpc/include/asm/kvm_ppc.h int (*get_one_reg)(struct kvm_vcpu *vcpu, u64 id, kvm_vcpu 270 arch/powerpc/include/asm/kvm_ppc.h int (*set_one_reg)(struct kvm_vcpu *vcpu, u64 id, kvm_vcpu 272 arch/powerpc/include/asm/kvm_ppc.h void (*vcpu_load)(struct kvm_vcpu *vcpu, int cpu); kvm_vcpu 273 arch/powerpc/include/asm/kvm_ppc.h void (*vcpu_put)(struct kvm_vcpu *vcpu); kvm_vcpu 274 arch/powerpc/include/asm/kvm_ppc.h void (*set_msr)(struct kvm_vcpu *vcpu, u64 msr); kvm_vcpu 275 arch/powerpc/include/asm/kvm_ppc.h int (*vcpu_run)(struct kvm_run *run, struct kvm_vcpu *vcpu); kvm_vcpu 276 arch/powerpc/include/asm/kvm_ppc.h struct kvm_vcpu *(*vcpu_create)(struct kvm *kvm, unsigned int id); kvm_vcpu 277 arch/powerpc/include/asm/kvm_ppc.h void (*vcpu_free)(struct kvm_vcpu *vcpu); kvm_vcpu 278 arch/powerpc/include/asm/kvm_ppc.h int (*check_requests)(struct kvm_vcpu *vcpu); kvm_vcpu 294 arch/powerpc/include/asm/kvm_ppc.h void (*mmu_destroy)(struct kvm_vcpu *vcpu); kvm_vcpu 302 arch/powerpc/include/asm/kvm_ppc.h int (*emulate_op)(struct kvm_run *run, struct kvm_vcpu *vcpu, kvm_vcpu 304 arch/powerpc/include/asm/kvm_ppc.h int (*emulate_mtspr)(struct kvm_vcpu *vcpu, int sprn, ulong spr_val); kvm_vcpu 305 arch/powerpc/include/asm/kvm_ppc.h int (*emulate_mfspr)(struct kvm_vcpu *vcpu, int sprn, ulong *spr_val); kvm_vcpu 306 arch/powerpc/include/asm/kvm_ppc.h void (*fast_vcpu_kick)(struct kvm_vcpu *vcpu); kvm_vcpu 318 arch/powerpc/include/asm/kvm_ppc.h void (*giveup_ext)(struct kvm_vcpu *vcpu, ulong msr); kvm_vcpu 320 arch/powerpc/include/asm/kvm_ppc.h int (*load_from_eaddr)(struct kvm_vcpu *vcpu, ulong *eaddr, void *ptr, kvm_vcpu 322 arch/powerpc/include/asm/kvm_ppc.h int (*store_to_eaddr)(struct kvm_vcpu *vcpu, ulong *eaddr, void *ptr, kvm_vcpu 329 arch/powerpc/include/asm/kvm_ppc.h static inline int kvmppc_get_last_inst(struct kvm_vcpu *vcpu, kvm_vcpu 416 arch/powerpc/include/asm/kvm_ppc.h int kvmppc_core_get_sregs(struct kvm_vcpu *vcpu, struct kvm_sregs *sregs); kvm_vcpu 417 arch/powerpc/include/asm/kvm_ppc.h int kvmppc_core_set_sregs(struct kvm_vcpu *vcpu, struct kvm_sregs *sregs); kvm_vcpu 419 arch/powerpc/include/asm/kvm_ppc.h int kvmppc_get_sregs_ivor(struct kvm_vcpu *vcpu, struct kvm_sregs *sregs); kvm_vcpu 420 arch/powerpc/include/asm/kvm_ppc.h int kvmppc_set_sregs_ivor(struct kvm_vcpu *vcpu, struct kvm_sregs *sregs); kvm_vcpu 422 arch/powerpc/include/asm/kvm_ppc.h int kvm_vcpu_ioctl_get_one_reg(struct kvm_vcpu *vcpu, struct kvm_one_reg *reg); kvm_vcpu 423 arch/powerpc/include/asm/kvm_ppc.h int kvm_vcpu_ioctl_set_one_reg(struct kvm_vcpu *vcpu, struct kvm_one_reg *reg); kvm_vcpu 424 arch/powerpc/include/asm/kvm_ppc.h int kvmppc_get_one_reg(struct kvm_vcpu *vcpu, u64 id, union kvmppc_one_reg *); kvm_vcpu 425 arch/powerpc/include/asm/kvm_ppc.h int kvmppc_set_one_reg(struct kvm_vcpu *vcpu, u64 id, union kvmppc_one_reg *); kvm_vcpu 427 arch/powerpc/include/asm/kvm_ppc.h void kvmppc_set_pid(struct kvm_vcpu *vcpu, u32 pid); kvm_vcpu 551 arch/powerpc/include/asm/kvm_ppc.h static inline void kvmppc_fast_vcpu_kick(struct kvm_vcpu *vcpu) kvm_vcpu 586 arch/powerpc/include/asm/kvm_ppc.h static inline void kvmppc_fast_vcpu_kick(struct kvm_vcpu *vcpu) kvm_vcpu 596 arch/powerpc/include/asm/kvm_ppc.h static inline int kvmppc_xics_enabled(struct kvm_vcpu *vcpu) kvm_vcpu 612 arch/powerpc/include/asm/kvm_ppc.h extern int kvmppc_xics_rm_complete(struct kvm_vcpu *vcpu, u32 hcall); kvm_vcpu 613 arch/powerpc/include/asm/kvm_ppc.h extern void kvmppc_xics_free_icp(struct kvm_vcpu *vcpu); kvm_vcpu 614 arch/powerpc/include/asm/kvm_ppc.h extern int kvmppc_xics_hcall(struct kvm_vcpu *vcpu, u32 cmd); kvm_vcpu 615 arch/powerpc/include/asm/kvm_ppc.h extern u64 kvmppc_xics_get_icp(struct kvm_vcpu *vcpu); kvm_vcpu 616 arch/powerpc/include/asm/kvm_ppc.h extern int kvmppc_xics_set_icp(struct kvm_vcpu *vcpu, u64 icpval); kvm_vcpu 618 arch/powerpc/include/asm/kvm_ppc.h struct kvm_vcpu *vcpu, u32 cpu); kvm_vcpu 624 arch/powerpc/include/asm/kvm_ppc.h extern long kvmppc_deliver_irq_passthru(struct kvm_vcpu *vcpu, __be32 xirr, kvm_vcpu 640 arch/powerpc/include/asm/kvm_ppc.h static inline int kvmppc_xics_rm_complete(struct kvm_vcpu *vcpu, u32 hcall) kvm_vcpu 642 arch/powerpc/include/asm/kvm_ppc.h static inline int kvmppc_xics_enabled(struct kvm_vcpu *vcpu) kvm_vcpu 644 arch/powerpc/include/asm/kvm_ppc.h static inline void kvmppc_xics_free_icp(struct kvm_vcpu *vcpu) { } kvm_vcpu 645 arch/powerpc/include/asm/kvm_ppc.h static inline int kvmppc_xics_hcall(struct kvm_vcpu *vcpu, u32 cmd) kvm_vcpu 668 arch/powerpc/include/asm/kvm_ppc.h struct kvm_vcpu *vcpu, u32 cpu); kvm_vcpu 669 arch/powerpc/include/asm/kvm_ppc.h extern void kvmppc_xive_cleanup_vcpu(struct kvm_vcpu *vcpu); kvm_vcpu 674 arch/powerpc/include/asm/kvm_ppc.h extern u64 kvmppc_xive_get_icp(struct kvm_vcpu *vcpu); kvm_vcpu 675 arch/powerpc/include/asm/kvm_ppc.h extern int kvmppc_xive_set_icp(struct kvm_vcpu *vcpu, u64 icpval); kvm_vcpu 679 arch/powerpc/include/asm/kvm_ppc.h extern void kvmppc_xive_push_vcpu(struct kvm_vcpu *vcpu); kvm_vcpu 681 arch/powerpc/include/asm/kvm_ppc.h static inline int kvmppc_xive_enabled(struct kvm_vcpu *vcpu) kvm_vcpu 687 arch/powerpc/include/asm/kvm_ppc.h struct kvm_vcpu *vcpu, u32 cpu); kvm_vcpu 688 arch/powerpc/include/asm/kvm_ppc.h extern void kvmppc_xive_native_cleanup_vcpu(struct kvm_vcpu *vcpu); kvm_vcpu 691 arch/powerpc/include/asm/kvm_ppc.h extern int kvmppc_xive_native_get_vp(struct kvm_vcpu *vcpu, kvm_vcpu 693 arch/powerpc/include/asm/kvm_ppc.h extern int kvmppc_xive_native_set_vp(struct kvm_vcpu *vcpu, kvm_vcpu 708 arch/powerpc/include/asm/kvm_ppc.h struct kvm_vcpu *vcpu, u32 cpu) { return -EBUSY; } kvm_vcpu 709 arch/powerpc/include/asm/kvm_ppc.h static inline void kvmppc_xive_cleanup_vcpu(struct kvm_vcpu *vcpu) { } kvm_vcpu 714 arch/powerpc/include/asm/kvm_ppc.h static inline u64 kvmppc_xive_get_icp(struct kvm_vcpu *vcpu) { return 0; } kvm_vcpu 715 arch/powerpc/include/asm/kvm_ppc.h static inline int kvmppc_xive_set_icp(struct kvm_vcpu *vcpu, u64 icpval) { return -ENOENT; } kvm_vcpu 719 arch/powerpc/include/asm/kvm_ppc.h static inline void kvmppc_xive_push_vcpu(struct kvm_vcpu *vcpu) { } kvm_vcpu 721 arch/powerpc/include/asm/kvm_ppc.h static inline int kvmppc_xive_enabled(struct kvm_vcpu *vcpu) kvm_vcpu 724 arch/powerpc/include/asm/kvm_ppc.h struct kvm_vcpu *vcpu, u32 cpu) { return -EBUSY; } kvm_vcpu 725 arch/powerpc/include/asm/kvm_ppc.h static inline void kvmppc_xive_native_cleanup_vcpu(struct kvm_vcpu *vcpu) { } kvm_vcpu 728 arch/powerpc/include/asm/kvm_ppc.h static inline int kvmppc_xive_native_get_vp(struct kvm_vcpu *vcpu, kvm_vcpu 731 arch/powerpc/include/asm/kvm_ppc.h static inline int kvmppc_xive_native_set_vp(struct kvm_vcpu *vcpu, kvm_vcpu 753 arch/powerpc/include/asm/kvm_ppc.h long kvmppc_rm_h_put_tce(struct kvm_vcpu *vcpu, unsigned long liobn, kvm_vcpu 755 arch/powerpc/include/asm/kvm_ppc.h long kvmppc_rm_h_put_tce_indirect(struct kvm_vcpu *vcpu, kvm_vcpu 758 arch/powerpc/include/asm/kvm_ppc.h long kvmppc_rm_h_stuff_tce(struct kvm_vcpu *vcpu, kvm_vcpu 761 arch/powerpc/include/asm/kvm_ppc.h long int kvmppc_rm_h_confer(struct kvm_vcpu *vcpu, int target, kvm_vcpu 763 arch/powerpc/include/asm/kvm_ppc.h long kvmppc_h_random(struct kvm_vcpu *vcpu); kvm_vcpu 765 arch/powerpc/include/asm/kvm_ppc.h void kvmppc_realmode_machine_check(struct kvm_vcpu *vcpu); kvm_vcpu 769 arch/powerpc/include/asm/kvm_ppc.h long kvmppc_h_enter(struct kvm_vcpu *vcpu, unsigned long flags, kvm_vcpu 771 arch/powerpc/include/asm/kvm_ppc.h long kvmppc_h_remove(struct kvm_vcpu *vcpu, unsigned long flags, kvm_vcpu 773 arch/powerpc/include/asm/kvm_ppc.h long kvmppc_h_bulk_remove(struct kvm_vcpu *vcpu); kvm_vcpu 774 arch/powerpc/include/asm/kvm_ppc.h long kvmppc_h_protect(struct kvm_vcpu *vcpu, unsigned long flags, kvm_vcpu 777 arch/powerpc/include/asm/kvm_ppc.h long kvmppc_h_read(struct kvm_vcpu *vcpu, unsigned long flags, kvm_vcpu 779 arch/powerpc/include/asm/kvm_ppc.h long kvmppc_h_clear_ref(struct kvm_vcpu *vcpu, unsigned long flags, kvm_vcpu 781 arch/powerpc/include/asm/kvm_ppc.h long kvmppc_h_clear_mod(struct kvm_vcpu *vcpu, unsigned long flags, kvm_vcpu 783 arch/powerpc/include/asm/kvm_ppc.h long kvmppc_rm_h_page_init(struct kvm_vcpu *vcpu, unsigned long flags, kvm_vcpu 785 arch/powerpc/include/asm/kvm_ppc.h long kvmppc_hpte_hv_fault(struct kvm_vcpu *vcpu, unsigned long addr, kvm_vcpu 787 arch/powerpc/include/asm/kvm_ppc.h unsigned long kvmppc_rm_h_xirr(struct kvm_vcpu *vcpu); kvm_vcpu 788 arch/powerpc/include/asm/kvm_ppc.h unsigned long kvmppc_rm_h_xirr_x(struct kvm_vcpu *vcpu); kvm_vcpu 789 arch/powerpc/include/asm/kvm_ppc.h unsigned long kvmppc_rm_h_ipoll(struct kvm_vcpu *vcpu, unsigned long server); kvm_vcpu 790 arch/powerpc/include/asm/kvm_ppc.h int kvmppc_rm_h_ipi(struct kvm_vcpu *vcpu, unsigned long server, kvm_vcpu 792 arch/powerpc/include/asm/kvm_ppc.h int kvmppc_rm_h_cppr(struct kvm_vcpu *vcpu, unsigned long cppr); kvm_vcpu 793 arch/powerpc/include/asm/kvm_ppc.h int kvmppc_rm_h_eoi(struct kvm_vcpu *vcpu, unsigned long xirr); kvm_vcpu 794 arch/powerpc/include/asm/kvm_ppc.h void kvmppc_guest_entry_inject_int(struct kvm_vcpu *vcpu); kvm_vcpu 818 arch/powerpc/include/asm/kvm_ppc.h void (*vcpu_kick)(struct kvm_vcpu *vcpu); kvm_vcpu 823 arch/powerpc/include/asm/kvm_ppc.h static inline unsigned long kvmppc_get_epr(struct kvm_vcpu *vcpu) kvm_vcpu 834 arch/powerpc/include/asm/kvm_ppc.h static inline void kvmppc_set_epr(struct kvm_vcpu *vcpu, u32 epr) kvm_vcpu 845 arch/powerpc/include/asm/kvm_ppc.h void kvmppc_mpic_set_epr(struct kvm_vcpu *vcpu); kvm_vcpu 846 arch/powerpc/include/asm/kvm_ppc.h int kvmppc_mpic_connect_vcpu(struct kvm_device *dev, struct kvm_vcpu *vcpu, kvm_vcpu 848 arch/powerpc/include/asm/kvm_ppc.h void kvmppc_mpic_disconnect_vcpu(struct openpic *opp, struct kvm_vcpu *vcpu); kvm_vcpu 852 arch/powerpc/include/asm/kvm_ppc.h static inline void kvmppc_mpic_set_epr(struct kvm_vcpu *vcpu) kvm_vcpu 857 arch/powerpc/include/asm/kvm_ppc.h struct kvm_vcpu *vcpu, u32 cpu) kvm_vcpu 863 arch/powerpc/include/asm/kvm_ppc.h struct kvm_vcpu *vcpu) kvm_vcpu 869 arch/powerpc/include/asm/kvm_ppc.h int kvm_vcpu_ioctl_config_tlb(struct kvm_vcpu *vcpu, kvm_vcpu 871 arch/powerpc/include/asm/kvm_ppc.h int kvm_vcpu_ioctl_dirty_tlb(struct kvm_vcpu *vcpu, kvm_vcpu 901 arch/powerpc/include/asm/kvm_ppc.h static inline bool kvmppc_shared_big_endian(struct kvm_vcpu *vcpu) kvm_vcpu 915 arch/powerpc/include/asm/kvm_ppc.h static inline ulong kvmppc_get_##reg(struct kvm_vcpu *vcpu) \ kvm_vcpu 921 arch/powerpc/include/asm/kvm_ppc.h static inline void kvmppc_set_##reg(struct kvm_vcpu *vcpu, ulong val) \ kvm_vcpu 927 arch/powerpc/include/asm/kvm_ppc.h static inline u##size kvmppc_get_##reg(struct kvm_vcpu *vcpu) \ kvm_vcpu 936 arch/powerpc/include/asm/kvm_ppc.h static inline void kvmppc_set_##reg(struct kvm_vcpu *vcpu, u##size val) \ kvm_vcpu 974 arch/powerpc/include/asm/kvm_ppc.h static inline void kvmppc_set_msr_fast(struct kvm_vcpu *vcpu, u64 val) kvm_vcpu 988 arch/powerpc/include/asm/kvm_ppc.h static inline u32 kvmppc_get_sr(struct kvm_vcpu *vcpu, int nr) kvm_vcpu 996 arch/powerpc/include/asm/kvm_ppc.h static inline void kvmppc_set_sr(struct kvm_vcpu *vcpu, int nr, u32 val) kvm_vcpu 1026 arch/powerpc/include/asm/kvm_ppc.h static inline ulong kvmppc_get_ea_indexed(struct kvm_vcpu *vcpu, int ra, int rb) kvm_vcpu 236 arch/powerpc/include/asm/processor.h struct kvm_vcpu *kvm_vcpu; kvm_vcpu 144 arch/powerpc/kernel/asm-offsets.c OFFSET(THREAD_KVM_VCPU, thread_struct, kvm_vcpu); kvm_vcpu 434 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_HOST_STACK, kvm_vcpu, arch.host_stack); kvm_vcpu 435 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_HOST_PID, kvm_vcpu, arch.host_pid); kvm_vcpu 436 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_GUEST_PID, kvm_vcpu, arch.pid); kvm_vcpu 437 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_GPRS, kvm_vcpu, arch.regs.gpr); kvm_vcpu 438 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_VRSAVE, kvm_vcpu, arch.vrsave); kvm_vcpu 439 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_FPRS, kvm_vcpu, arch.fp.fpr); kvm_vcpu 441 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_VRS, kvm_vcpu, arch.vr.vr); kvm_vcpu 443 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_XER, kvm_vcpu, arch.regs.xer); kvm_vcpu 444 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_CTR, kvm_vcpu, arch.regs.ctr); kvm_vcpu 445 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_LR, kvm_vcpu, arch.regs.link); kvm_vcpu 447 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_TAR, kvm_vcpu, arch.tar); kvm_vcpu 449 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_CR, kvm_vcpu, arch.regs.ccr); kvm_vcpu 450 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_PC, kvm_vcpu, arch.regs.nip); kvm_vcpu 452 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_MSR, kvm_vcpu, arch.shregs.msr); kvm_vcpu 453 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_SRR0, kvm_vcpu, arch.shregs.srr0); kvm_vcpu 454 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_SRR1, kvm_vcpu, arch.shregs.srr1); kvm_vcpu 455 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_SPRG0, kvm_vcpu, arch.shregs.sprg0); kvm_vcpu 456 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_SPRG1, kvm_vcpu, arch.shregs.sprg1); kvm_vcpu 457 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_SPRG2, kvm_vcpu, arch.shregs.sprg2); kvm_vcpu 458 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_SPRG3, kvm_vcpu, arch.shregs.sprg3); kvm_vcpu 461 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_TB_RMENTRY, kvm_vcpu, arch.rm_entry); kvm_vcpu 462 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_TB_RMINTR, kvm_vcpu, arch.rm_intr); kvm_vcpu 463 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_TB_RMEXIT, kvm_vcpu, arch.rm_exit); kvm_vcpu 464 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_TB_GUEST, kvm_vcpu, arch.guest_time); kvm_vcpu 465 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_TB_CEDE, kvm_vcpu, arch.cede_time); kvm_vcpu 466 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_CUR_ACTIVITY, kvm_vcpu, arch.cur_activity); kvm_vcpu 467 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_ACTIVITY_START, kvm_vcpu, arch.cur_tb_start); kvm_vcpu 478 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_SHADOW_PID, kvm_vcpu, arch.shadow_pid); kvm_vcpu 479 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_SHADOW_PID1, kvm_vcpu, arch.shadow_pid1); kvm_vcpu 480 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_SHARED, kvm_vcpu, arch.shared); kvm_vcpu 482 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_SHADOW_MSR, kvm_vcpu, arch.shadow_msr); kvm_vcpu 484 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_SHAREDBE, kvm_vcpu, arch.shared_big_endian); kvm_vcpu 494 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_KVM, kvm_vcpu, kvm); kvm_vcpu 510 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_DSISR, kvm_vcpu, arch.shregs.dsisr); kvm_vcpu 511 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_DAR, kvm_vcpu, arch.shregs.dar); kvm_vcpu 512 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_VPA, kvm_vcpu, arch.vpa.pinned_addr); kvm_vcpu 513 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_VPA_DIRTY, kvm_vcpu, arch.vpa.dirty); kvm_vcpu 514 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_HEIR, kvm_vcpu, arch.emul_inst); kvm_vcpu 515 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_NESTED, kvm_vcpu, arch.nested); kvm_vcpu 516 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_CPU, kvm_vcpu, cpu); kvm_vcpu 517 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_THREAD_CPU, kvm_vcpu, arch.thread_cpu); kvm_vcpu 520 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_PURR, kvm_vcpu, arch.purr); kvm_vcpu 521 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_SPURR, kvm_vcpu, arch.spurr); kvm_vcpu 522 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_IC, kvm_vcpu, arch.ic); kvm_vcpu 523 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_DSCR, kvm_vcpu, arch.dscr); kvm_vcpu 524 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_AMR, kvm_vcpu, arch.amr); kvm_vcpu 525 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_UAMOR, kvm_vcpu, arch.uamor); kvm_vcpu 526 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_IAMR, kvm_vcpu, arch.iamr); kvm_vcpu 527 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_CTRL, kvm_vcpu, arch.ctrl); kvm_vcpu 528 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_DABR, kvm_vcpu, arch.dabr); kvm_vcpu 529 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_DABRX, kvm_vcpu, arch.dabrx); kvm_vcpu 530 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_DAWR, kvm_vcpu, arch.dawr); kvm_vcpu 531 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_DAWRX, kvm_vcpu, arch.dawrx); kvm_vcpu 532 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_CIABR, kvm_vcpu, arch.ciabr); kvm_vcpu 533 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_HFLAGS, kvm_vcpu, arch.hflags); kvm_vcpu 534 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_DEC, kvm_vcpu, arch.dec); kvm_vcpu 535 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_DEC_EXPIRES, kvm_vcpu, arch.dec_expires); kvm_vcpu 536 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_PENDING_EXC, kvm_vcpu, arch.pending_exceptions); kvm_vcpu 537 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_CEDED, kvm_vcpu, arch.ceded); kvm_vcpu 538 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_PRODDED, kvm_vcpu, arch.prodded); kvm_vcpu 539 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_IRQ_PENDING, kvm_vcpu, arch.irq_pending); kvm_vcpu 540 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_DBELL_REQ, kvm_vcpu, arch.doorbell_request); kvm_vcpu 541 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_MMCR, kvm_vcpu, arch.mmcr); kvm_vcpu 542 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_PMC, kvm_vcpu, arch.pmc); kvm_vcpu 543 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_SPMC, kvm_vcpu, arch.spmc); kvm_vcpu 544 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_SIAR, kvm_vcpu, arch.siar); kvm_vcpu 545 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_SDAR, kvm_vcpu, arch.sdar); kvm_vcpu 546 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_SIER, kvm_vcpu, arch.sier); kvm_vcpu 547 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_SLB, kvm_vcpu, arch.slb); kvm_vcpu 548 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_SLB_MAX, kvm_vcpu, arch.slb_max); kvm_vcpu 549 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_SLB_NR, kvm_vcpu, arch.slb_nr); kvm_vcpu 550 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_FAULT_DSISR, kvm_vcpu, arch.fault_dsisr); kvm_vcpu 551 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_FAULT_DAR, kvm_vcpu, arch.fault_dar); kvm_vcpu 552 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_FAULT_GPA, kvm_vcpu, arch.fault_gpa); kvm_vcpu 553 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_INTR_MSR, kvm_vcpu, arch.intr_msr); kvm_vcpu 554 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_LAST_INST, kvm_vcpu, arch.last_inst); kvm_vcpu 555 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_TRAP, kvm_vcpu, arch.trap); kvm_vcpu 556 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_CFAR, kvm_vcpu, arch.cfar); kvm_vcpu 557 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_PPR, kvm_vcpu, arch.ppr); kvm_vcpu 558 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_FSCR, kvm_vcpu, arch.fscr); kvm_vcpu 559 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_PSPB, kvm_vcpu, arch.pspb); kvm_vcpu 560 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_EBBHR, kvm_vcpu, arch.ebbhr); kvm_vcpu 561 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_EBBRR, kvm_vcpu, arch.ebbrr); kvm_vcpu 562 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_BESCR, kvm_vcpu, arch.bescr); kvm_vcpu 563 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_CSIGR, kvm_vcpu, arch.csigr); kvm_vcpu 564 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_TACR, kvm_vcpu, arch.tacr); kvm_vcpu 565 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_TCSCR, kvm_vcpu, arch.tcscr); kvm_vcpu 566 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_ACOP, kvm_vcpu, arch.acop); kvm_vcpu 567 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_WORT, kvm_vcpu, arch.wort); kvm_vcpu 568 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_TID, kvm_vcpu, arch.tid); kvm_vcpu 569 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_PSSCR, kvm_vcpu, arch.psscr); kvm_vcpu 570 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_HFSCR, kvm_vcpu, arch.hfscr); kvm_vcpu 585 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_TFHAR, kvm_vcpu, arch.tfhar); kvm_vcpu 586 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_TFIAR, kvm_vcpu, arch.tfiar); kvm_vcpu 587 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_TEXASR, kvm_vcpu, arch.texasr); kvm_vcpu 588 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_ORIG_TEXASR, kvm_vcpu, arch.orig_texasr); kvm_vcpu 589 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_GPR_TM, kvm_vcpu, arch.gpr_tm); kvm_vcpu 590 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_FPRS_TM, kvm_vcpu, arch.fp_tm.fpr); kvm_vcpu 591 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_VRS_TM, kvm_vcpu, arch.vr_tm.vr); kvm_vcpu 592 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_VRSAVE_TM, kvm_vcpu, arch.vrsave_tm); kvm_vcpu 593 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_CR_TM, kvm_vcpu, arch.cr_tm); kvm_vcpu 594 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_XER_TM, kvm_vcpu, arch.xer_tm); kvm_vcpu 595 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_LR_TM, kvm_vcpu, arch.lr_tm); kvm_vcpu 596 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_CTR_TM, kvm_vcpu, arch.ctr_tm); kvm_vcpu 597 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_AMR_TM, kvm_vcpu, arch.amr_tm); kvm_vcpu 598 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_PPR_TM, kvm_vcpu, arch.ppr_tm); kvm_vcpu 599 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_DSCR_TM, kvm_vcpu, arch.dscr_tm); kvm_vcpu 600 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_TAR_TM, kvm_vcpu, arch.tar_tm); kvm_vcpu 662 arch/powerpc/kernel/asm-offsets.c HSTATE_FIELD(HSTATE_KVM_VCPU, kvm_vcpu); kvm_vcpu 708 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_CR, kvm_vcpu, arch.regs.ccr); kvm_vcpu 709 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_XER, kvm_vcpu, arch.regs.xer); kvm_vcpu 710 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_LR, kvm_vcpu, arch.regs.link); kvm_vcpu 711 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_CTR, kvm_vcpu, arch.regs.ctr); kvm_vcpu 712 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_PC, kvm_vcpu, arch.regs.nip); kvm_vcpu 713 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_SPRG9, kvm_vcpu, arch.sprg9); kvm_vcpu 714 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_LAST_INST, kvm_vcpu, arch.last_inst); kvm_vcpu 715 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_FAULT_DEAR, kvm_vcpu, arch.fault_dear); kvm_vcpu 716 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_FAULT_ESR, kvm_vcpu, arch.fault_esr); kvm_vcpu 717 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_CRIT_SAVE, kvm_vcpu, arch.crit_save); kvm_vcpu 745 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_EVR, kvm_vcpu, arch.evr[0]); kvm_vcpu 746 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_ACC, kvm_vcpu, arch.acc); kvm_vcpu 747 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_SPEFSCR, kvm_vcpu, arch.spefscr); kvm_vcpu 748 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_HOST_SPEFSCR, kvm_vcpu, arch.host_spefscr); kvm_vcpu 752 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_HOST_MAS4, kvm_vcpu, arch.host_mas4); kvm_vcpu 753 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_HOST_MAS6, kvm_vcpu, arch.host_mas6); kvm_vcpu 757 arch/powerpc/kernel/asm-offsets.c DEFINE(VCPU_XIVE_SAVED_STATE, offsetof(struct kvm_vcpu, kvm_vcpu 759 arch/powerpc/kernel/asm-offsets.c DEFINE(VCPU_XIVE_CAM_WORD, offsetof(struct kvm_vcpu, kvm_vcpu 761 arch/powerpc/kernel/asm-offsets.c DEFINE(VCPU_XIVE_PUSHED, offsetof(struct kvm_vcpu, arch.xive_pushed)); kvm_vcpu 762 arch/powerpc/kernel/asm-offsets.c DEFINE(VCPU_XIVE_ESC_ON, offsetof(struct kvm_vcpu, arch.xive_esc_on)); kvm_vcpu 763 arch/powerpc/kernel/asm-offsets.c DEFINE(VCPU_XIVE_ESC_RADDR, offsetof(struct kvm_vcpu, arch.xive_esc_raddr)); kvm_vcpu 764 arch/powerpc/kernel/asm-offsets.c DEFINE(VCPU_XIVE_ESC_VADDR, offsetof(struct kvm_vcpu, arch.xive_esc_vaddr)); kvm_vcpu 768 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_TIMING_EXIT_TBU, kvm_vcpu, arch.timing_exit.tv32.tbu); kvm_vcpu 769 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_TIMING_EXIT_TBL, kvm_vcpu, arch.timing_exit.tv32.tbl); kvm_vcpu 770 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_TIMING_LAST_ENTER_TBU, kvm_vcpu, arch.timing_last_enter.tv32.tbu); kvm_vcpu 771 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_TIMING_LAST_ENTER_TBL, kvm_vcpu, arch.timing_last_enter.tv32.tbl); kvm_vcpu 40 arch/powerpc/kvm/book3s.c #define VCPU_STAT(x, ...) offsetof(struct kvm_vcpu, stat.x), KVM_STAT_VCPU, ## __VA_ARGS__ kvm_vcpu 77 arch/powerpc/kvm/book3s.c void kvmppc_unfixup_split_real(struct kvm_vcpu *vcpu) kvm_vcpu 91 arch/powerpc/kvm/book3s.c static inline unsigned long kvmppc_interrupt_offset(struct kvm_vcpu *vcpu) kvm_vcpu 98 arch/powerpc/kvm/book3s.c static inline void kvmppc_update_int_pending(struct kvm_vcpu *vcpu, kvm_vcpu 109 arch/powerpc/kvm/book3s.c static inline bool kvmppc_critical_section(struct kvm_vcpu *vcpu) kvm_vcpu 135 arch/powerpc/kvm/book3s.c void kvmppc_inject_interrupt(struct kvm_vcpu *vcpu, int vec, u64 flags) kvm_vcpu 171 arch/powerpc/kvm/book3s.c void kvmppc_book3s_dequeue_irqprio(struct kvm_vcpu *vcpu, kvm_vcpu 183 arch/powerpc/kvm/book3s.c void kvmppc_book3s_queue_irqprio(struct kvm_vcpu *vcpu, unsigned int vec) kvm_vcpu 195 arch/powerpc/kvm/book3s.c void kvmppc_core_queue_machine_check(struct kvm_vcpu *vcpu, ulong flags) kvm_vcpu 202 arch/powerpc/kvm/book3s.c void kvmppc_core_queue_program(struct kvm_vcpu *vcpu, ulong flags) kvm_vcpu 209 arch/powerpc/kvm/book3s.c void kvmppc_core_queue_fpunavail(struct kvm_vcpu *vcpu) kvm_vcpu 215 arch/powerpc/kvm/book3s.c void kvmppc_core_queue_vec_unavail(struct kvm_vcpu *vcpu) kvm_vcpu 221 arch/powerpc/kvm/book3s.c void kvmppc_core_queue_vsx_unavail(struct kvm_vcpu *vcpu) kvm_vcpu 227 arch/powerpc/kvm/book3s.c void kvmppc_core_queue_dec(struct kvm_vcpu *vcpu) kvm_vcpu 233 arch/powerpc/kvm/book3s.c int kvmppc_core_pending_dec(struct kvm_vcpu *vcpu) kvm_vcpu 239 arch/powerpc/kvm/book3s.c void kvmppc_core_dequeue_dec(struct kvm_vcpu *vcpu) kvm_vcpu 245 arch/powerpc/kvm/book3s.c void kvmppc_core_queue_external(struct kvm_vcpu *vcpu, kvm_vcpu 274 arch/powerpc/kvm/book3s.c void kvmppc_core_dequeue_external(struct kvm_vcpu *vcpu) kvm_vcpu 279 arch/powerpc/kvm/book3s.c void kvmppc_core_queue_data_storage(struct kvm_vcpu *vcpu, ulong dar, kvm_vcpu 288 arch/powerpc/kvm/book3s.c void kvmppc_core_queue_inst_storage(struct kvm_vcpu *vcpu, ulong flags) kvm_vcpu 294 arch/powerpc/kvm/book3s.c static int kvmppc_book3s_irqprio_deliver(struct kvm_vcpu *vcpu, kvm_vcpu 374 arch/powerpc/kvm/book3s.c static bool clear_irqprio(struct kvm_vcpu *vcpu, unsigned int priority) kvm_vcpu 396 arch/powerpc/kvm/book3s.c int kvmppc_core_prepare_to_enter(struct kvm_vcpu *vcpu) kvm_vcpu 426 arch/powerpc/kvm/book3s.c kvm_pfn_t kvmppc_gpa_to_pfn(struct kvm_vcpu *vcpu, gpa_t gpa, bool writing, kvm_vcpu 452 arch/powerpc/kvm/book3s.c int kvmppc_xlate(struct kvm_vcpu *vcpu, ulong eaddr, enum xlate_instdata xlid, kvm_vcpu 482 arch/powerpc/kvm/book3s.c int kvmppc_load_last_inst(struct kvm_vcpu *vcpu, kvm_vcpu 499 arch/powerpc/kvm/book3s.c int kvm_arch_vcpu_setup(struct kvm_vcpu *vcpu) kvm_vcpu 504 arch/powerpc/kvm/book3s.c int kvmppc_subarch_vcpu_init(struct kvm_vcpu *vcpu) kvm_vcpu 509 arch/powerpc/kvm/book3s.c void kvmppc_subarch_vcpu_uninit(struct kvm_vcpu *vcpu) kvm_vcpu 513 arch/powerpc/kvm/book3s.c int kvm_arch_vcpu_ioctl_get_sregs(struct kvm_vcpu *vcpu, kvm_vcpu 525 arch/powerpc/kvm/book3s.c int kvm_arch_vcpu_ioctl_set_sregs(struct kvm_vcpu *vcpu, kvm_vcpu 537 arch/powerpc/kvm/book3s.c int kvm_arch_vcpu_ioctl_get_regs(struct kvm_vcpu *vcpu, struct kvm_regs *regs) kvm_vcpu 565 arch/powerpc/kvm/book3s.c int kvm_arch_vcpu_ioctl_set_regs(struct kvm_vcpu *vcpu, struct kvm_regs *regs) kvm_vcpu 592 arch/powerpc/kvm/book3s.c int kvm_arch_vcpu_ioctl_get_fpu(struct kvm_vcpu *vcpu, struct kvm_fpu *fpu) kvm_vcpu 597 arch/powerpc/kvm/book3s.c int kvm_arch_vcpu_ioctl_set_fpu(struct kvm_vcpu *vcpu, struct kvm_fpu *fpu) kvm_vcpu 602 arch/powerpc/kvm/book3s.c int kvmppc_get_one_reg(struct kvm_vcpu *vcpu, u64 id, kvm_vcpu 690 arch/powerpc/kvm/book3s.c int kvmppc_set_one_reg(struct kvm_vcpu *vcpu, u64 id, kvm_vcpu 775 arch/powerpc/kvm/book3s.c void kvmppc_core_vcpu_load(struct kvm_vcpu *vcpu, int cpu) kvm_vcpu 780 arch/powerpc/kvm/book3s.c void kvmppc_core_vcpu_put(struct kvm_vcpu *vcpu) kvm_vcpu 785 arch/powerpc/kvm/book3s.c void kvmppc_set_msr(struct kvm_vcpu *vcpu, u64 msr) kvm_vcpu 791 arch/powerpc/kvm/book3s.c int kvmppc_vcpu_run(struct kvm_run *kvm_run, struct kvm_vcpu *vcpu) kvm_vcpu 796 arch/powerpc/kvm/book3s.c int kvm_arch_vcpu_ioctl_translate(struct kvm_vcpu *vcpu, kvm_vcpu 802 arch/powerpc/kvm/book3s.c int kvm_arch_vcpu_ioctl_set_guest_debug(struct kvm_vcpu *vcpu, kvm_vcpu 811 arch/powerpc/kvm/book3s.c void kvmppc_decrementer_func(struct kvm_vcpu *vcpu) kvm_vcpu 817 arch/powerpc/kvm/book3s.c struct kvm_vcpu *kvmppc_core_vcpu_create(struct kvm *kvm, unsigned int id) kvm_vcpu 822 arch/powerpc/kvm/book3s.c void kvmppc_core_vcpu_free(struct kvm_vcpu *vcpu) kvm_vcpu 827 arch/powerpc/kvm/book3s.c int kvmppc_core_check_requests(struct kvm_vcpu *vcpu) kvm_vcpu 891 arch/powerpc/kvm/book3s.c void kvmppc_mmu_destroy(struct kvm_vcpu *vcpu) kvm_vcpu 929 arch/powerpc/kvm/book3s.c int kvmppc_h_logical_ci_load(struct kvm_vcpu *vcpu) kvm_vcpu 971 arch/powerpc/kvm/book3s.c int kvmppc_h_logical_ci_store(struct kvm_vcpu *vcpu) kvm_vcpu 1074 arch/powerpc/kvm/book3s.c r = kvm_init(NULL, sizeof(struct kvm_vcpu), 0, THIS_MODULE); kvm_vcpu 19 arch/powerpc/kvm/book3s.h extern void kvmppc_mmu_destroy_pr(struct kvm_vcpu *vcpu); kvm_vcpu 20 arch/powerpc/kvm/book3s.h extern int kvmppc_core_emulate_op_pr(struct kvm_run *run, struct kvm_vcpu *vcpu, kvm_vcpu 22 arch/powerpc/kvm/book3s.h extern int kvmppc_core_emulate_mtspr_pr(struct kvm_vcpu *vcpu, kvm_vcpu 24 arch/powerpc/kvm/book3s.h extern int kvmppc_core_emulate_mfspr_pr(struct kvm_vcpu *vcpu, kvm_vcpu 30 arch/powerpc/kvm/book3s.h extern void kvmppc_emulate_tabort(struct kvm_vcpu *vcpu, int ra_val); kvm_vcpu 32 arch/powerpc/kvm/book3s.h static inline void kvmppc_emulate_tabort(struct kvm_vcpu *vcpu, int ra_val) {} kvm_vcpu 40 arch/powerpc/kvm/book3s_32_mmu.c static inline bool check_debug_ip(struct kvm_vcpu *vcpu) kvm_vcpu 69 arch/powerpc/kvm/book3s_32_mmu.c static int kvmppc_mmu_book3s_32_xlate_bat(struct kvm_vcpu *vcpu, gva_t eaddr, kvm_vcpu 72 arch/powerpc/kvm/book3s_32_mmu.c static int kvmppc_mmu_book3s_32_esid_to_vsid(struct kvm_vcpu *vcpu, ulong esid, kvm_vcpu 75 arch/powerpc/kvm/book3s_32_mmu.c static u32 find_sr(struct kvm_vcpu *vcpu, gva_t eaddr) kvm_vcpu 80 arch/powerpc/kvm/book3s_32_mmu.c static u64 kvmppc_mmu_book3s_32_ea_to_vp(struct kvm_vcpu *vcpu, gva_t eaddr, kvm_vcpu 93 arch/powerpc/kvm/book3s_32_mmu.c static void kvmppc_mmu_book3s_32_reset_msr(struct kvm_vcpu *vcpu) kvm_vcpu 98 arch/powerpc/kvm/book3s_32_mmu.c static hva_t kvmppc_mmu_book3s_32_get_pteg(struct kvm_vcpu *vcpu, kvm_vcpu 132 arch/powerpc/kvm/book3s_32_mmu.c static int kvmppc_mmu_book3s_32_xlate_bat(struct kvm_vcpu *vcpu, gva_t eaddr, kvm_vcpu 187 arch/powerpc/kvm/book3s_32_mmu.c static int kvmppc_mmu_book3s_32_xlate_pte(struct kvm_vcpu *vcpu, gva_t eaddr, kvm_vcpu 295 arch/powerpc/kvm/book3s_32_mmu.c static int kvmppc_mmu_book3s_32_xlate(struct kvm_vcpu *vcpu, gva_t eaddr, kvm_vcpu 331 arch/powerpc/kvm/book3s_32_mmu.c static u32 kvmppc_mmu_book3s_32_mfsrin(struct kvm_vcpu *vcpu, u32 srnum) kvm_vcpu 336 arch/powerpc/kvm/book3s_32_mmu.c static void kvmppc_mmu_book3s_32_mtsrin(struct kvm_vcpu *vcpu, u32 srnum, kvm_vcpu 343 arch/powerpc/kvm/book3s_32_mmu.c static void kvmppc_mmu_book3s_32_tlbie(struct kvm_vcpu *vcpu, ulong ea, bool large) kvm_vcpu 346 arch/powerpc/kvm/book3s_32_mmu.c struct kvm_vcpu *v; kvm_vcpu 353 arch/powerpc/kvm/book3s_32_mmu.c static int kvmppc_mmu_book3s_32_esid_to_vsid(struct kvm_vcpu *vcpu, ulong esid, kvm_vcpu 396 arch/powerpc/kvm/book3s_32_mmu.c static bool kvmppc_mmu_book3s_32_is_dcbz32(struct kvm_vcpu *vcpu) kvm_vcpu 402 arch/powerpc/kvm/book3s_32_mmu.c void kvmppc_mmu_book3s_32_init(struct kvm_vcpu *vcpu) kvm_vcpu 49 arch/powerpc/kvm/book3s_32_mmu_host.c void kvmppc_mmu_invalidate_pte(struct kvm_vcpu *vcpu, struct hpte_cache *pte) kvm_vcpu 66 arch/powerpc/kvm/book3s_32_mmu_host.c static u16 kvmppc_sid_hash(struct kvm_vcpu *vcpu, u64 gvsid) kvm_vcpu 79 arch/powerpc/kvm/book3s_32_mmu_host.c static struct kvmppc_sid_map *find_sid_vsid(struct kvm_vcpu *vcpu, u64 gvsid) kvm_vcpu 106 arch/powerpc/kvm/book3s_32_mmu_host.c static u32 *kvmppc_mmu_get_pteg(struct kvm_vcpu *vcpu, u32 vsid, u32 eaddr, kvm_vcpu 130 arch/powerpc/kvm/book3s_32_mmu_host.c int kvmppc_mmu_map_page(struct kvm_vcpu *vcpu, struct kvmppc_pte *orig_pte, kvm_vcpu 258 arch/powerpc/kvm/book3s_32_mmu_host.c void kvmppc_mmu_unmap_page(struct kvm_vcpu *vcpu, struct kvmppc_pte *pte) kvm_vcpu 263 arch/powerpc/kvm/book3s_32_mmu_host.c static struct kvmppc_sid_map *create_sid_map(struct kvm_vcpu *vcpu, u64 gvsid) kvm_vcpu 302 arch/powerpc/kvm/book3s_32_mmu_host.c int kvmppc_mmu_map_segment(struct kvm_vcpu *vcpu, ulong eaddr) kvm_vcpu 333 arch/powerpc/kvm/book3s_32_mmu_host.c void kvmppc_mmu_flush_segments(struct kvm_vcpu *vcpu) kvm_vcpu 345 arch/powerpc/kvm/book3s_32_mmu_host.c void kvmppc_mmu_destroy_pr(struct kvm_vcpu *vcpu) kvm_vcpu 359 arch/powerpc/kvm/book3s_32_mmu_host.c int kvmppc_mmu_init(struct kvm_vcpu *vcpu) kvm_vcpu 27 arch/powerpc/kvm/book3s_64_mmu.c static void kvmppc_mmu_book3s_64_reset_msr(struct kvm_vcpu *vcpu) kvm_vcpu 42 arch/powerpc/kvm/book3s_64_mmu.c struct kvm_vcpu *vcpu, kvm_vcpu 95 arch/powerpc/kvm/book3s_64_mmu.c static u64 kvmppc_mmu_book3s_64_ea_to_vp(struct kvm_vcpu *vcpu, gva_t eaddr, kvm_vcpu 130 arch/powerpc/kvm/book3s_64_mmu.c static hva_t kvmppc_mmu_book3s_64_get_pteg(struct kvm_vcpu *vcpu, kvm_vcpu 205 arch/powerpc/kvm/book3s_64_mmu.c static int kvmppc_mmu_book3s_64_xlate(struct kvm_vcpu *vcpu, gva_t eaddr, kvm_vcpu 378 arch/powerpc/kvm/book3s_64_mmu.c static void kvmppc_mmu_book3s_64_slbmte(struct kvm_vcpu *vcpu, u64 rs, u64 rb) kvm_vcpu 427 arch/powerpc/kvm/book3s_64_mmu.c static int kvmppc_mmu_book3s_64_slbfee(struct kvm_vcpu *vcpu, gva_t eaddr, kvm_vcpu 440 arch/powerpc/kvm/book3s_64_mmu.c static u64 kvmppc_mmu_book3s_64_slbmfee(struct kvm_vcpu *vcpu, u64 slb_nr) kvm_vcpu 452 arch/powerpc/kvm/book3s_64_mmu.c static u64 kvmppc_mmu_book3s_64_slbmfev(struct kvm_vcpu *vcpu, u64 slb_nr) kvm_vcpu 464 arch/powerpc/kvm/book3s_64_mmu.c static void kvmppc_mmu_book3s_64_slbie(struct kvm_vcpu *vcpu, u64 ea) kvm_vcpu 486 arch/powerpc/kvm/book3s_64_mmu.c static void kvmppc_mmu_book3s_64_slbia(struct kvm_vcpu *vcpu) kvm_vcpu 504 arch/powerpc/kvm/book3s_64_mmu.c static void kvmppc_mmu_book3s_64_mtsrin(struct kvm_vcpu *vcpu, u32 srnum, kvm_vcpu 544 arch/powerpc/kvm/book3s_64_mmu.c static void kvmppc_mmu_book3s_64_tlbie(struct kvm_vcpu *vcpu, ulong va, kvm_vcpu 549 arch/powerpc/kvm/book3s_64_mmu.c struct kvm_vcpu *v; kvm_vcpu 578 arch/powerpc/kvm/book3s_64_mmu.c static int segment_contains_magic_page(struct kvm_vcpu *vcpu, ulong esid) kvm_vcpu 587 arch/powerpc/kvm/book3s_64_mmu.c static int kvmppc_mmu_book3s_64_esid_to_vsid(struct kvm_vcpu *vcpu, ulong esid, kvm_vcpu 661 arch/powerpc/kvm/book3s_64_mmu.c static bool kvmppc_mmu_book3s_64_is_dcbz32(struct kvm_vcpu *vcpu) kvm_vcpu 666 arch/powerpc/kvm/book3s_64_mmu.c void kvmppc_mmu_book3s_64_init(struct kvm_vcpu *vcpu) kvm_vcpu 23 arch/powerpc/kvm/book3s_64_mmu_host.c void kvmppc_mmu_invalidate_pte(struct kvm_vcpu *vcpu, struct hpte_cache *pte) kvm_vcpu 32 arch/powerpc/kvm/book3s_64_mmu_host.c static u16 kvmppc_sid_hash(struct kvm_vcpu *vcpu, u64 gvsid) kvm_vcpu 45 arch/powerpc/kvm/book3s_64_mmu_host.c static struct kvmppc_sid_map *find_sid_vsid(struct kvm_vcpu *vcpu, u64 gvsid) kvm_vcpu 70 arch/powerpc/kvm/book3s_64_mmu_host.c int kvmppc_mmu_map_page(struct kvm_vcpu *vcpu, struct kvmppc_pte *orig_pte, kvm_vcpu 212 arch/powerpc/kvm/book3s_64_mmu_host.c void kvmppc_mmu_unmap_page(struct kvm_vcpu *vcpu, struct kvmppc_pte *pte) kvm_vcpu 223 arch/powerpc/kvm/book3s_64_mmu_host.c static struct kvmppc_sid_map *create_sid_map(struct kvm_vcpu *vcpu, u64 gvsid) kvm_vcpu 269 arch/powerpc/kvm/book3s_64_mmu_host.c static int kvmppc_mmu_next_segment(struct kvm_vcpu *vcpu, ulong esid) kvm_vcpu 310 arch/powerpc/kvm/book3s_64_mmu_host.c int kvmppc_mmu_map_segment(struct kvm_vcpu *vcpu, ulong eaddr) kvm_vcpu 356 arch/powerpc/kvm/book3s_64_mmu_host.c void kvmppc_mmu_flush_segment(struct kvm_vcpu *vcpu, ulong ea, ulong seg_size) kvm_vcpu 373 arch/powerpc/kvm/book3s_64_mmu_host.c void kvmppc_mmu_flush_segments(struct kvm_vcpu *vcpu) kvm_vcpu 381 arch/powerpc/kvm/book3s_64_mmu_host.c void kvmppc_mmu_destroy_pr(struct kvm_vcpu *vcpu) kvm_vcpu 387 arch/powerpc/kvm/book3s_64_mmu_host.c int kvmppc_mmu_init(struct kvm_vcpu *vcpu) kvm_vcpu 204 arch/powerpc/kvm/book3s_64_mmu_hv.c void kvmppc_map_vrma(struct kvm_vcpu *vcpu, struct kvm_memory_slot *memslot, kvm_vcpu 278 arch/powerpc/kvm/book3s_64_mmu_hv.c static void kvmppc_mmu_book3s_64_hv_reset_msr(struct kvm_vcpu *vcpu) kvm_vcpu 310 arch/powerpc/kvm/book3s_64_mmu_hv.c static struct kvmppc_slb *kvmppc_mmu_book3s_hv_find_slbe(struct kvm_vcpu *vcpu, kvm_vcpu 340 arch/powerpc/kvm/book3s_64_mmu_hv.c static int kvmppc_mmu_book3s_64_hv_xlate(struct kvm_vcpu *vcpu, gva_t eaddr, kvm_vcpu 428 arch/powerpc/kvm/book3s_64_mmu_hv.c int kvmppc_hv_emulate_mmio(struct kvm_run *run, struct kvm_vcpu *vcpu, kvm_vcpu 491 arch/powerpc/kvm/book3s_64_mmu_hv.c int kvmppc_book3s_hv_page_fault(struct kvm_run *run, struct kvm_vcpu *vcpu, kvm_vcpu 2157 arch/powerpc/kvm/book3s_64_mmu_hv.c void kvmppc_mmu_book3s_hv_init(struct kvm_vcpu *vcpu) kvm_vcpu 84 arch/powerpc/kvm/book3s_64_mmu_radix.c static long kvmhv_copy_tofrom_guest_radix(struct kvm_vcpu *vcpu, gva_t eaddr, kvm_vcpu 107 arch/powerpc/kvm/book3s_64_mmu_radix.c long kvmhv_copy_from_guest_radix(struct kvm_vcpu *vcpu, gva_t eaddr, void *to, kvm_vcpu 120 arch/powerpc/kvm/book3s_64_mmu_radix.c long kvmhv_copy_to_guest_radix(struct kvm_vcpu *vcpu, gva_t eaddr, void *from, kvm_vcpu 127 arch/powerpc/kvm/book3s_64_mmu_radix.c int kvmppc_mmu_walk_radix_tree(struct kvm_vcpu *vcpu, gva_t eaddr, kvm_vcpu 220 arch/powerpc/kvm/book3s_64_mmu_radix.c int kvmppc_mmu_radix_translate_table(struct kvm_vcpu *vcpu, gva_t eaddr, kvm_vcpu 249 arch/powerpc/kvm/book3s_64_mmu_radix.c int kvmppc_mmu_radix_xlate(struct kvm_vcpu *vcpu, gva_t eaddr, kvm_vcpu 766 arch/powerpc/kvm/book3s_64_mmu_radix.c int kvmppc_book3s_instantiate_page(struct kvm_vcpu *vcpu, kvm_vcpu 889 arch/powerpc/kvm/book3s_64_mmu_radix.c int kvmppc_book3s_radix_page_fault(struct kvm_run *run, struct kvm_vcpu *vcpu, kvm_vcpu 540 arch/powerpc/kvm/book3s_64_vio.c long kvmppc_h_put_tce(struct kvm_vcpu *vcpu, unsigned long liobn, kvm_vcpu 600 arch/powerpc/kvm/book3s_64_vio.c long kvmppc_h_put_tce_indirect(struct kvm_vcpu *vcpu, kvm_vcpu 697 arch/powerpc/kvm/book3s_64_vio.c long kvmppc_h_stuff_tce(struct kvm_vcpu *vcpu, kvm_vcpu 386 arch/powerpc/kvm/book3s_64_vio_hv.c long kvmppc_rm_h_put_tce(struct kvm_vcpu *vcpu, unsigned long liobn, kvm_vcpu 441 arch/powerpc/kvm/book3s_64_vio_hv.c static long kvmppc_rm_ua_to_hpa(struct kvm_vcpu *vcpu, kvm_vcpu 477 arch/powerpc/kvm/book3s_64_vio_hv.c long kvmppc_rm_h_put_tce_indirect(struct kvm_vcpu *vcpu, kvm_vcpu 600 arch/powerpc/kvm/book3s_64_vio_hv.c long kvmppc_rm_h_stuff_tce(struct kvm_vcpu *vcpu, kvm_vcpu 653 arch/powerpc/kvm/book3s_64_vio_hv.c long kvmppc_h_get_tce(struct kvm_vcpu *vcpu, unsigned long liobn, kvm_vcpu 74 arch/powerpc/kvm/book3s_emulate.c static bool spr_allowed(struct kvm_vcpu *vcpu, enum priv_level level) kvm_vcpu 88 arch/powerpc/kvm/book3s_emulate.c static inline void kvmppc_copyto_vcpu_tm(struct kvm_vcpu *vcpu) kvm_vcpu 107 arch/powerpc/kvm/book3s_emulate.c static inline void kvmppc_copyfrom_vcpu_tm(struct kvm_vcpu *vcpu) kvm_vcpu 126 arch/powerpc/kvm/book3s_emulate.c static void kvmppc_emulate_treclaim(struct kvm_vcpu *vcpu, int ra_val) kvm_vcpu 172 arch/powerpc/kvm/book3s_emulate.c static void kvmppc_emulate_trchkpt(struct kvm_vcpu *vcpu) kvm_vcpu 197 arch/powerpc/kvm/book3s_emulate.c void kvmppc_emulate_tabort(struct kvm_vcpu *vcpu, int ra_val) kvm_vcpu 238 arch/powerpc/kvm/book3s_emulate.c int kvmppc_core_emulate_op_pr(struct kvm_run *run, struct kvm_vcpu *vcpu, kvm_vcpu 637 arch/powerpc/kvm/book3s_emulate.c void kvmppc_set_bat(struct kvm_vcpu *vcpu, struct kvmppc_bat *bat, bool upper, kvm_vcpu 657 arch/powerpc/kvm/book3s_emulate.c static struct kvmppc_bat *kvmppc_find_bat(struct kvm_vcpu *vcpu, int sprn) kvm_vcpu 682 arch/powerpc/kvm/book3s_emulate.c int kvmppc_core_emulate_mtspr_pr(struct kvm_vcpu *vcpu, int sprn, ulong spr_val) kvm_vcpu 865 arch/powerpc/kvm/book3s_emulate.c int kvmppc_core_emulate_mfspr_pr(struct kvm_vcpu *vcpu, int sprn, ulong *spr_val) kvm_vcpu 1032 arch/powerpc/kvm/book3s_emulate.c u32 kvmppc_alignment_dsisr(struct kvm_vcpu *vcpu, unsigned int inst) kvm_vcpu 1037 arch/powerpc/kvm/book3s_emulate.c ulong kvmppc_alignment_dar(struct kvm_vcpu *vcpu, unsigned int inst) kvm_vcpu 136 arch/powerpc/kvm/book3s_hv.c static void kvmppc_end_cede(struct kvm_vcpu *vcpu); kvm_vcpu 137 arch/powerpc/kvm/book3s_hv.c static int kvmppc_hv_setup_htab_rma(struct kvm_vcpu *vcpu); kvm_vcpu 165 arch/powerpc/kvm/book3s_hv.c static inline struct kvm_vcpu *next_runnable_thread(struct kvmppc_vcore *vc, kvm_vcpu 169 arch/powerpc/kvm/book3s_hv.c struct kvm_vcpu *vcpu; kvm_vcpu 229 arch/powerpc/kvm/book3s_hv.c static void kvmppc_fast_vcpu_kick_hv(struct kvm_vcpu *vcpu) kvm_vcpu 304 arch/powerpc/kvm/book3s_hv.c static void kvmppc_core_vcpu_load_hv(struct kvm_vcpu *vcpu, int cpu) kvm_vcpu 327 arch/powerpc/kvm/book3s_hv.c static void kvmppc_core_vcpu_put_hv(struct kvm_vcpu *vcpu) kvm_vcpu 341 arch/powerpc/kvm/book3s_hv.c static void kvmppc_set_msr_hv(struct kvm_vcpu *vcpu, u64 msr) kvm_vcpu 353 arch/powerpc/kvm/book3s_hv.c static void kvmppc_set_pvr_hv(struct kvm_vcpu *vcpu, u32 pvr) kvm_vcpu 361 arch/powerpc/kvm/book3s_hv.c static int kvmppc_set_arch_compat(struct kvm_vcpu *vcpu, u32 arch_compat) kvm_vcpu 414 arch/powerpc/kvm/book3s_hv.c static void kvmppc_dump_regs(struct kvm_vcpu *vcpu) kvm_vcpu 447 arch/powerpc/kvm/book3s_hv.c static struct kvm_vcpu *kvmppc_find_vcpu(struct kvm *kvm, int id) kvm_vcpu 452 arch/powerpc/kvm/book3s_hv.c static void init_vpa(struct kvm_vcpu *vcpu, struct lppaca *vpa) kvm_vcpu 458 arch/powerpc/kvm/book3s_hv.c static int set_vpa(struct kvm_vcpu *vcpu, struct kvmppc_vpa *v, kvm_vcpu 490 arch/powerpc/kvm/book3s_hv.c static unsigned long do_h_register_vpa(struct kvm_vcpu *vcpu, kvm_vcpu 497 arch/powerpc/kvm/book3s_hv.c struct kvm_vcpu *tvcpu; kvm_vcpu 607 arch/powerpc/kvm/book3s_hv.c static void kvmppc_update_vpa(struct kvm_vcpu *vcpu, struct kvmppc_vpa *vpap) kvm_vcpu 657 arch/powerpc/kvm/book3s_hv.c static void kvmppc_update_vpas(struct kvm_vcpu *vcpu) kvm_vcpu 698 arch/powerpc/kvm/book3s_hv.c static void kvmppc_create_dtl_entry(struct kvm_vcpu *vcpu, kvm_vcpu 738 arch/powerpc/kvm/book3s_hv.c static bool kvmppc_doorbell_pending(struct kvm_vcpu *vcpu) kvm_vcpu 756 arch/powerpc/kvm/book3s_hv.c static bool kvmppc_power8_compatible(struct kvm_vcpu *vcpu) kvm_vcpu 766 arch/powerpc/kvm/book3s_hv.c static int kvmppc_h_set_mode(struct kvm_vcpu *vcpu, unsigned long mflags, kvm_vcpu 842 arch/powerpc/kvm/book3s_hv.c static long kvmppc_h_page_init(struct kvm_vcpu *vcpu, unsigned long flags, kvm_vcpu 874 arch/powerpc/kvm/book3s_hv.c static int kvm_arch_vcpu_yield_to(struct kvm_vcpu *target) kvm_vcpu 896 arch/powerpc/kvm/book3s_hv.c static int kvmppc_get_yield_count(struct kvm_vcpu *vcpu) kvm_vcpu 909 arch/powerpc/kvm/book3s_hv.c int kvmppc_pseries_do_hcall(struct kvm_vcpu *vcpu) kvm_vcpu 914 arch/powerpc/kvm/book3s_hv.c struct kvm_vcpu *tvcpu; kvm_vcpu 1095 arch/powerpc/kvm/book3s_hv.c static void kvmppc_nested_cede(struct kvm_vcpu *vcpu) kvm_vcpu 1134 arch/powerpc/kvm/book3s_hv.c struct kvm_vcpu *vcpu) kvm_vcpu 1161 arch/powerpc/kvm/book3s_hv.c static unsigned long kvmppc_read_dpdes(struct kvm_vcpu *vcpu) kvm_vcpu 1164 arch/powerpc/kvm/book3s_hv.c struct kvm_vcpu *v; kvm_vcpu 1194 arch/powerpc/kvm/book3s_hv.c static int kvmppc_emulate_doorbell_instr(struct kvm_vcpu *vcpu) kvm_vcpu 1199 arch/powerpc/kvm/book3s_hv.c struct kvm_vcpu *tvcpu; kvm_vcpu 1250 arch/powerpc/kvm/book3s_hv.c static int kvmppc_handle_exit_hv(struct kvm_run *run, struct kvm_vcpu *vcpu, kvm_vcpu 1439 arch/powerpc/kvm/book3s_hv.c static int kvmppc_handle_nested_exit(struct kvm_run *run, struct kvm_vcpu *vcpu) kvm_vcpu 1537 arch/powerpc/kvm/book3s_hv.c static int kvm_arch_vcpu_ioctl_get_sregs_hv(struct kvm_vcpu *vcpu, kvm_vcpu 1552 arch/powerpc/kvm/book3s_hv.c static int kvm_arch_vcpu_ioctl_set_sregs_hv(struct kvm_vcpu *vcpu, kvm_vcpu 1574 arch/powerpc/kvm/book3s_hv.c static void kvmppc_set_lpcr(struct kvm_vcpu *vcpu, u64 new_lpcr, kvm_vcpu 1587 arch/powerpc/kvm/book3s_hv.c struct kvm_vcpu *vcpu; kvm_vcpu 1622 arch/powerpc/kvm/book3s_hv.c static int kvmppc_get_one_reg_hv(struct kvm_vcpu *vcpu, u64 id, kvm_vcpu 1843 arch/powerpc/kvm/book3s_hv.c static int kvmppc_set_one_reg_hv(struct kvm_vcpu *vcpu, u64 id, kvm_vcpu 2117 arch/powerpc/kvm/book3s_hv.c {"rm_entry", offsetof(struct kvm_vcpu, arch.rm_entry)}, kvm_vcpu 2118 arch/powerpc/kvm/book3s_hv.c {"rm_intr", offsetof(struct kvm_vcpu, arch.rm_intr)}, kvm_vcpu 2119 arch/powerpc/kvm/book3s_hv.c {"rm_exit", offsetof(struct kvm_vcpu, arch.rm_exit)}, kvm_vcpu 2120 arch/powerpc/kvm/book3s_hv.c {"guest", offsetof(struct kvm_vcpu, arch.guest_time)}, kvm_vcpu 2121 arch/powerpc/kvm/book3s_hv.c {"cede", offsetof(struct kvm_vcpu, arch.cede_time)}, kvm_vcpu 2127 arch/powerpc/kvm/book3s_hv.c struct kvm_vcpu *vcpu; kvm_vcpu 2134 arch/powerpc/kvm/book3s_hv.c struct kvm_vcpu *vcpu = inode->i_private; kvm_vcpu 2161 arch/powerpc/kvm/book3s_hv.c struct kvm_vcpu *vcpu = p->vcpu; kvm_vcpu 2238 arch/powerpc/kvm/book3s_hv.c static void debugfs_vcpu_init(struct kvm_vcpu *vcpu, unsigned int id) kvm_vcpu 2255 arch/powerpc/kvm/book3s_hv.c static void debugfs_vcpu_init(struct kvm_vcpu *vcpu, unsigned int id) kvm_vcpu 2260 arch/powerpc/kvm/book3s_hv.c static struct kvm_vcpu *kvmppc_core_vcpu_create_hv(struct kvm *kvm, kvm_vcpu 2263 arch/powerpc/kvm/book3s_hv.c struct kvm_vcpu *vcpu; kvm_vcpu 2426 arch/powerpc/kvm/book3s_hv.c static void kvmppc_core_vcpu_free_hv(struct kvm_vcpu *vcpu) kvm_vcpu 2437 arch/powerpc/kvm/book3s_hv.c static int kvmppc_core_check_requests_hv(struct kvm_vcpu *vcpu) kvm_vcpu 2443 arch/powerpc/kvm/book3s_hv.c static void kvmppc_set_timer(struct kvm_vcpu *vcpu) kvm_vcpu 2459 arch/powerpc/kvm/book3s_hv.c static void kvmppc_end_cede(struct kvm_vcpu *vcpu) kvm_vcpu 2471 arch/powerpc/kvm/book3s_hv.c struct kvm_vcpu *vcpu) kvm_vcpu 2496 arch/powerpc/kvm/book3s_hv.c tpaca->kvm_hstate.kvm_vcpu = NULL; kvm_vcpu 2528 arch/powerpc/kvm/book3s_hv.c tpaca->kvm_hstate.kvm_vcpu = NULL; kvm_vcpu 2533 arch/powerpc/kvm/book3s_hv.c static void radix_flush_cpu(struct kvm *kvm, int cpu, struct kvm_vcpu *vcpu) kvm_vcpu 2558 arch/powerpc/kvm/book3s_hv.c static void kvmppc_prepare_radix_vcpu(struct kvm_vcpu *vcpu, int pcpu) kvm_vcpu 2596 arch/powerpc/kvm/book3s_hv.c static void kvmppc_start_thread(struct kvm_vcpu *vcpu, struct kvmppc_vcore *vc) kvm_vcpu 2614 arch/powerpc/kvm/book3s_hv.c tpaca->kvm_hstate.kvm_vcpu = vcpu; kvm_vcpu 2849 arch/powerpc/kvm/book3s_hv.c struct kvm_vcpu *vcpu; kvm_vcpu 2899 arch/powerpc/kvm/book3s_hv.c struct kvm_vcpu *vcpu; kvm_vcpu 2918 arch/powerpc/kvm/book3s_hv.c struct kvm_vcpu *vcpu; kvm_vcpu 3044 arch/powerpc/kvm/book3s_hv.c struct kvm_vcpu *vcpu; kvm_vcpu 3389 arch/powerpc/kvm/book3s_hv.c static int kvmhv_load_hv_regs_and_go(struct kvm_vcpu *vcpu, u64 time_limit, kvm_vcpu 3519 arch/powerpc/kvm/book3s_hv.c int kvmhv_p9_guest_entry(struct kvm_vcpu *vcpu, u64 time_limit, kvm_vcpu 3709 arch/powerpc/kvm/book3s_hv.c struct kvm_vcpu *vcpu, int wait_state) kvm_vcpu 3741 arch/powerpc/kvm/book3s_hv.c static inline bool xive_interrupt_pending(struct kvm_vcpu *vcpu) kvm_vcpu 3749 arch/powerpc/kvm/book3s_hv.c static inline bool xive_interrupt_pending(struct kvm_vcpu *vcpu) kvm_vcpu 3755 arch/powerpc/kvm/book3s_hv.c static bool kvmppc_vcpu_woken(struct kvm_vcpu *vcpu) kvm_vcpu 3770 arch/powerpc/kvm/book3s_hv.c struct kvm_vcpu *vcpu; kvm_vcpu 3887 arch/powerpc/kvm/book3s_hv.c static int kvmhv_setup_mmu(struct kvm_vcpu *vcpu) kvm_vcpu 3906 arch/powerpc/kvm/book3s_hv.c static int kvmppc_run_vcpu(struct kvm_run *kvm_run, struct kvm_vcpu *vcpu) kvm_vcpu 3910 arch/powerpc/kvm/book3s_hv.c struct kvm_vcpu *v; kvm_vcpu 4037 arch/powerpc/kvm/book3s_hv.c struct kvm_vcpu *vcpu, u64 time_limit, kvm_vcpu 4215 arch/powerpc/kvm/book3s_hv.c static int kvmppc_vcpu_run_hv(struct kvm_run *run, struct kvm_vcpu *vcpu) kvm_vcpu 4400 arch/powerpc/kvm/book3s_hv.c struct kvm_vcpu *vcpu; kvm_vcpu 4546 arch/powerpc/kvm/book3s_hv.c static void kvmppc_mmu_destroy_hv(struct kvm_vcpu *vcpu) kvm_vcpu 4576 arch/powerpc/kvm/book3s_hv.c static int kvmppc_hv_setup_htab_rma(struct kvm_vcpu *vcpu) kvm_vcpu 4967 arch/powerpc/kvm/book3s_hv.c static int kvmppc_core_emulate_op_hv(struct kvm_run *run, struct kvm_vcpu *vcpu, kvm_vcpu 4973 arch/powerpc/kvm/book3s_hv.c static int kvmppc_core_emulate_mtspr_hv(struct kvm_vcpu *vcpu, int sprn, kvm_vcpu 4979 arch/powerpc/kvm/book3s_hv.c static int kvmppc_core_emulate_mfspr_hv(struct kvm_vcpu *vcpu, int sprn, kvm_vcpu 5362 arch/powerpc/kvm/book3s_hv.c static int kvmhv_load_from_eaddr(struct kvm_vcpu *vcpu, ulong *eaddr, void *ptr, kvm_vcpu 5381 arch/powerpc/kvm/book3s_hv.c static int kvmhv_store_to_eaddr(struct kvm_vcpu *vcpu, ulong *eaddr, void *ptr, kvm_vcpu 39 arch/powerpc/kvm/book3s_hv_builtin.c unsigned long (*__xive_vm_h_xirr)(struct kvm_vcpu *vcpu); kvm_vcpu 40 arch/powerpc/kvm/book3s_hv_builtin.c unsigned long (*__xive_vm_h_ipoll)(struct kvm_vcpu *vcpu, unsigned long server); kvm_vcpu 41 arch/powerpc/kvm/book3s_hv_builtin.c int (*__xive_vm_h_ipi)(struct kvm_vcpu *vcpu, unsigned long server, kvm_vcpu 43 arch/powerpc/kvm/book3s_hv_builtin.c int (*__xive_vm_h_cppr)(struct kvm_vcpu *vcpu, unsigned long cppr); kvm_vcpu 44 arch/powerpc/kvm/book3s_hv_builtin.c int (*__xive_vm_h_eoi)(struct kvm_vcpu *vcpu, unsigned long xirr); kvm_vcpu 132 arch/powerpc/kvm/book3s_hv_builtin.c long int kvmppc_rm_h_confer(struct kvm_vcpu *vcpu, int target, kvm_vcpu 206 arch/powerpc/kvm/book3s_hv_builtin.c long kvmppc_h_random(struct kvm_vcpu *vcpu) kvm_vcpu 397 arch/powerpc/kvm/book3s_hv_builtin.c struct kvm_vcpu *vcpu; kvm_vcpu 399 arch/powerpc/kvm/book3s_hv_builtin.c vcpu = local_paca->kvm_hstate.kvm_vcpu; kvm_vcpu 573 arch/powerpc/kvm/book3s_hv_builtin.c unsigned long kvmppc_rm_h_xirr(struct kvm_vcpu *vcpu) kvm_vcpu 587 arch/powerpc/kvm/book3s_hv_builtin.c unsigned long kvmppc_rm_h_xirr_x(struct kvm_vcpu *vcpu) kvm_vcpu 602 arch/powerpc/kvm/book3s_hv_builtin.c unsigned long kvmppc_rm_h_ipoll(struct kvm_vcpu *vcpu, unsigned long server) kvm_vcpu 616 arch/powerpc/kvm/book3s_hv_builtin.c int kvmppc_rm_h_ipi(struct kvm_vcpu *vcpu, unsigned long server, kvm_vcpu 631 arch/powerpc/kvm/book3s_hv_builtin.c int kvmppc_rm_h_cppr(struct kvm_vcpu *vcpu, unsigned long cppr) kvm_vcpu 645 arch/powerpc/kvm/book3s_hv_builtin.c int kvmppc_rm_h_eoi(struct kvm_vcpu *vcpu, unsigned long xirr) kvm_vcpu 762 arch/powerpc/kvm/book3s_hv_builtin.c void kvmppc_guest_entry_inject_int(struct kvm_vcpu *vcpu) kvm_vcpu 28 arch/powerpc/kvm/book3s_hv_nested.c void kvmhv_save_hv_regs(struct kvm_vcpu *vcpu, struct hv_guest_state *hr) kvm_vcpu 96 arch/powerpc/kvm/book3s_hv_nested.c static void save_hv_return_state(struct kvm_vcpu *vcpu, int trap, kvm_vcpu 131 arch/powerpc/kvm/book3s_hv_nested.c static void sanitise_hv_regs(struct kvm_vcpu *vcpu, struct hv_guest_state *hr) kvm_vcpu 147 arch/powerpc/kvm/book3s_hv_nested.c static void restore_hv_regs(struct kvm_vcpu *vcpu, struct hv_guest_state *hr) kvm_vcpu 172 arch/powerpc/kvm/book3s_hv_nested.c void kvmhv_restore_hv_return_state(struct kvm_vcpu *vcpu, kvm_vcpu 198 arch/powerpc/kvm/book3s_hv_nested.c static void kvmhv_nested_mmio_needed(struct kvm_vcpu *vcpu, u64 regs_ptr) kvm_vcpu 218 arch/powerpc/kvm/book3s_hv_nested.c long kvmhv_enter_nested_guest(struct kvm_vcpu *vcpu) kvm_vcpu 443 arch/powerpc/kvm/book3s_hv_nested.c long kvmhv_set_partition_table(struct kvm_vcpu *vcpu) kvm_vcpu 473 arch/powerpc/kvm/book3s_hv_nested.c long kvmhv_copy_tofrom_guest_nested(struct kvm_vcpu *vcpu) kvm_vcpu 914 arch/powerpc/kvm/book3s_hv_nested.c static bool kvmhv_invalidate_shadow_pte(struct kvm_vcpu *vcpu, kvm_vcpu 973 arch/powerpc/kvm/book3s_hv_nested.c static int kvmhv_emulate_tlbie_tlb_addr(struct kvm_vcpu *vcpu, int lpid, kvm_vcpu 1009 arch/powerpc/kvm/book3s_hv_nested.c static void kvmhv_emulate_tlbie_lpid(struct kvm_vcpu *vcpu, kvm_vcpu 1040 arch/powerpc/kvm/book3s_hv_nested.c static void kvmhv_emulate_tlbie_all_lpid(struct kvm_vcpu *vcpu, int ric) kvm_vcpu 1058 arch/powerpc/kvm/book3s_hv_nested.c static int kvmhv_emulate_priv_tlbie(struct kvm_vcpu *vcpu, unsigned int instr, kvm_vcpu 1121 arch/powerpc/kvm/book3s_hv_nested.c long kvmhv_do_nested_tlbie(struct kvm_vcpu *vcpu) kvm_vcpu 1133 arch/powerpc/kvm/book3s_hv_nested.c static int kvmhv_translate_addr_nested(struct kvm_vcpu *vcpu, kvm_vcpu 1195 arch/powerpc/kvm/book3s_hv_nested.c static long kvmhv_handle_nested_set_rc(struct kvm_vcpu *vcpu, kvm_vcpu 1261 arch/powerpc/kvm/book3s_hv_nested.c struct kvm_vcpu *vcpu, kvm_vcpu 1431 arch/powerpc/kvm/book3s_hv_nested.c long int kvmhv_nested_page_fault(struct kvm_run *run, struct kvm_vcpu *vcpu) kvm_vcpu 36 arch/powerpc/kvm/book3s_hv_ras.c static void reload_slb(struct kvm_vcpu *vcpu) kvm_vcpu 68 arch/powerpc/kvm/book3s_hv_ras.c static void kvmppc_realmode_mc_power7(struct kvm_vcpu *vcpu) kvm_vcpu 125 arch/powerpc/kvm/book3s_hv_ras.c void kvmppc_realmode_machine_check(struct kvm_vcpu *vcpu) kvm_vcpu 55 arch/powerpc/kvm/book3s_hv_rm_mmu.c if (kvm->arch.online_vcores == 1 && local_paca->kvm_hstate.kvm_vcpu) kvm_vcpu 415 arch/powerpc/kvm/book3s_hv_rm_mmu.c long kvmppc_h_enter(struct kvm_vcpu *vcpu, unsigned long flags, kvm_vcpu 560 arch/powerpc/kvm/book3s_hv_rm_mmu.c long kvmppc_h_remove(struct kvm_vcpu *vcpu, unsigned long flags, kvm_vcpu 567 arch/powerpc/kvm/book3s_hv_rm_mmu.c long kvmppc_h_bulk_remove(struct kvm_vcpu *vcpu) kvm_vcpu 685 arch/powerpc/kvm/book3s_hv_rm_mmu.c long kvmppc_h_protect(struct kvm_vcpu *vcpu, unsigned long flags, kvm_vcpu 756 arch/powerpc/kvm/book3s_hv_rm_mmu.c long kvmppc_h_read(struct kvm_vcpu *vcpu, unsigned long flags, kvm_vcpu 796 arch/powerpc/kvm/book3s_hv_rm_mmu.c long kvmppc_h_clear_ref(struct kvm_vcpu *vcpu, unsigned long flags, kvm_vcpu 844 arch/powerpc/kvm/book3s_hv_rm_mmu.c long kvmppc_h_clear_mod(struct kvm_vcpu *vcpu, unsigned long flags, kvm_vcpu 891 arch/powerpc/kvm/book3s_hv_rm_mmu.c static int kvmppc_get_hpa(struct kvm_vcpu *vcpu, unsigned long gpa, kvm_vcpu 933 arch/powerpc/kvm/book3s_hv_rm_mmu.c static long kvmppc_do_h_page_init_zero(struct kvm_vcpu *vcpu, kvm_vcpu 967 arch/powerpc/kvm/book3s_hv_rm_mmu.c static long kvmppc_do_h_page_init_copy(struct kvm_vcpu *vcpu, kvm_vcpu 1003 arch/powerpc/kvm/book3s_hv_rm_mmu.c long kvmppc_rm_h_page_init(struct kvm_vcpu *vcpu, unsigned long flags, kvm_vcpu 1080 arch/powerpc/kvm/book3s_hv_rm_mmu.c static struct mmio_hpte_cache_entry *mmio_cache_search(struct kvm_vcpu *vcpu, kvm_vcpu 1100 arch/powerpc/kvm/book3s_hv_rm_mmu.c next_mmio_cache_entry(struct kvm_vcpu *vcpu) kvm_vcpu 1208 arch/powerpc/kvm/book3s_hv_rm_mmu.c long kvmppc_hpte_hv_fault(struct kvm_vcpu *vcpu, unsigned long addr, kvm_vcpu 54 arch/powerpc/kvm/book3s_hv_rm_xics.c static inline void icp_send_hcore_msg(int hcore, struct kvm_vcpu *vcpu) kvm_vcpu 66 arch/powerpc/kvm/book3s_hv_rm_xics.c static inline void icp_send_hcore_msg(int hcore, struct kvm_vcpu *vcpu) { } kvm_vcpu 127 arch/powerpc/kvm/book3s_hv_rm_xics.c static void icp_rm_set_vcpu_irq(struct kvm_vcpu *vcpu, kvm_vcpu 128 arch/powerpc/kvm/book3s_hv_rm_xics.c struct kvm_vcpu *this_vcpu) kvm_vcpu 174 arch/powerpc/kvm/book3s_hv_rm_xics.c static void icp_rm_clr_vcpu_irq(struct kvm_vcpu *vcpu) kvm_vcpu 185 arch/powerpc/kvm/book3s_hv_rm_xics.c struct kvm_vcpu *this_vcpu = local_paca->kvm_hstate.kvm_vcpu; kvm_vcpu 490 arch/powerpc/kvm/book3s_hv_rm_xics.c unsigned long xics_rm_h_xirr(struct kvm_vcpu *vcpu) kvm_vcpu 528 arch/powerpc/kvm/book3s_hv_rm_xics.c int xics_rm_h_ipi(struct kvm_vcpu *vcpu, unsigned long server, kvm_vcpu 615 arch/powerpc/kvm/book3s_hv_rm_xics.c int xics_rm_h_cppr(struct kvm_vcpu *vcpu, unsigned long cppr) kvm_vcpu 677 arch/powerpc/kvm/book3s_hv_rm_xics.c static int ics_rm_eoi(struct kvm_vcpu *vcpu, u32 irq) kvm_vcpu 735 arch/powerpc/kvm/book3s_hv_rm_xics.c int xics_rm_h_eoi(struct kvm_vcpu *vcpu, unsigned long xirr) kvm_vcpu 853 arch/powerpc/kvm/book3s_hv_rm_xics.c long kvmppc_deliver_irq_passthru(struct kvm_vcpu *vcpu, kvm_vcpu 14 arch/powerpc/kvm/book3s_hv_tm.c static void emulate_tx_failure(struct kvm_vcpu *vcpu, u64 failure_cause) kvm_vcpu 40 arch/powerpc/kvm/book3s_hv_tm.c int kvmhv_p9_tm_emulation(struct kvm_vcpu *vcpu) kvm_vcpu 20 arch/powerpc/kvm/book3s_hv_tm_builtin.c int kvmhv_p9_tm_emulation_early(struct kvm_vcpu *vcpu) kvm_vcpu 101 arch/powerpc/kvm/book3s_hv_tm_builtin.c void kvmhv_emulate_tm_rollback(struct kvm_vcpu *vcpu) kvm_vcpu 56 arch/powerpc/kvm/book3s_mmu_hpte.c void kvmppc_mmu_hpte_cache_map(struct kvm_vcpu *vcpu, struct hpte_cache *pte) kvm_vcpu 101 arch/powerpc/kvm/book3s_mmu_hpte.c static void invalidate_pte(struct kvm_vcpu *vcpu, struct hpte_cache *pte) kvm_vcpu 132 arch/powerpc/kvm/book3s_mmu_hpte.c static void kvmppc_mmu_pte_flush_all(struct kvm_vcpu *vcpu) kvm_vcpu 150 arch/powerpc/kvm/book3s_mmu_hpte.c static void kvmppc_mmu_pte_flush_page(struct kvm_vcpu *vcpu, ulong guest_ea) kvm_vcpu 169 arch/powerpc/kvm/book3s_mmu_hpte.c static void kvmppc_mmu_pte_flush_long(struct kvm_vcpu *vcpu, ulong guest_ea) kvm_vcpu 189 arch/powerpc/kvm/book3s_mmu_hpte.c void kvmppc_mmu_pte_flush(struct kvm_vcpu *vcpu, ulong guest_ea, ulong ea_mask) kvm_vcpu 212 arch/powerpc/kvm/book3s_mmu_hpte.c static void kvmppc_mmu_pte_vflush_short(struct kvm_vcpu *vcpu, u64 guest_vp) kvm_vcpu 233 arch/powerpc/kvm/book3s_mmu_hpte.c static void kvmppc_mmu_pte_vflush_64k(struct kvm_vcpu *vcpu, u64 guest_vp) kvm_vcpu 255 arch/powerpc/kvm/book3s_mmu_hpte.c static void kvmppc_mmu_pte_vflush_long(struct kvm_vcpu *vcpu, u64 guest_vp) kvm_vcpu 275 arch/powerpc/kvm/book3s_mmu_hpte.c void kvmppc_mmu_pte_vflush(struct kvm_vcpu *vcpu, u64 guest_vp, u64 vp_mask) kvm_vcpu 298 arch/powerpc/kvm/book3s_mmu_hpte.c void kvmppc_mmu_pte_pflush(struct kvm_vcpu *vcpu, ulong pa_start, ulong pa_end) kvm_vcpu 320 arch/powerpc/kvm/book3s_mmu_hpte.c struct hpte_cache *kvmppc_mmu_hpte_cache_next(struct kvm_vcpu *vcpu) kvm_vcpu 338 arch/powerpc/kvm/book3s_mmu_hpte.c void kvmppc_mmu_hpte_destroy(struct kvm_vcpu *vcpu) kvm_vcpu 351 arch/powerpc/kvm/book3s_mmu_hpte.c int kvmppc_mmu_hpte_init(struct kvm_vcpu *vcpu) kvm_vcpu 150 arch/powerpc/kvm/book3s_paired_singles.c static inline void kvmppc_sync_qpr(struct kvm_vcpu *vcpu, int rt) kvm_vcpu 155 arch/powerpc/kvm/book3s_paired_singles.c static void kvmppc_inject_pf(struct kvm_vcpu *vcpu, ulong eaddr, bool is_store) kvm_vcpu 172 arch/powerpc/kvm/book3s_paired_singles.c static int kvmppc_emulate_fpr_load(struct kvm_run *run, struct kvm_vcpu *vcpu, kvm_vcpu 216 arch/powerpc/kvm/book3s_paired_singles.c static int kvmppc_emulate_fpr_store(struct kvm_run *run, struct kvm_vcpu *vcpu, kvm_vcpu 262 arch/powerpc/kvm/book3s_paired_singles.c static int kvmppc_emulate_psq_load(struct kvm_run *run, struct kvm_vcpu *vcpu, kvm_vcpu 305 arch/powerpc/kvm/book3s_paired_singles.c static int kvmppc_emulate_psq_store(struct kvm_run *run, struct kvm_vcpu *vcpu, kvm_vcpu 344 arch/powerpc/kvm/book3s_paired_singles.c static bool kvmppc_inst_is_paired_single(struct kvm_vcpu *vcpu, u32 inst) kvm_vcpu 490 arch/powerpc/kvm/book3s_paired_singles.c static int kvmppc_ps_three_in(struct kvm_vcpu *vcpu, bool rc, kvm_vcpu 538 arch/powerpc/kvm/book3s_paired_singles.c static int kvmppc_ps_two_in(struct kvm_vcpu *vcpu, bool rc, kvm_vcpu 590 arch/powerpc/kvm/book3s_paired_singles.c static int kvmppc_ps_one_in(struct kvm_vcpu *vcpu, bool rc, kvm_vcpu 621 arch/powerpc/kvm/book3s_paired_singles.c int kvmppc_emulate_paired_single(struct kvm_run *run, struct kvm_vcpu *vcpu) kvm_vcpu 52 arch/powerpc/kvm/book3s_pr.c static int kvmppc_handle_ext(struct kvm_vcpu *vcpu, unsigned int exit_nr, kvm_vcpu 55 arch/powerpc/kvm/book3s_pr.c static int kvmppc_handle_fac(struct kvm_vcpu *vcpu, ulong fac); kvm_vcpu 66 arch/powerpc/kvm/book3s_pr.c static bool kvmppc_is_split_real(struct kvm_vcpu *vcpu) kvm_vcpu 72 arch/powerpc/kvm/book3s_pr.c static void kvmppc_fixup_split_real(struct kvm_vcpu *vcpu) kvm_vcpu 93 arch/powerpc/kvm/book3s_pr.c void kvmppc_unfixup_split_real(struct kvm_vcpu *vcpu); kvm_vcpu 95 arch/powerpc/kvm/book3s_pr.c static void kvmppc_core_vcpu_load_pr(struct kvm_vcpu *vcpu, int cpu) kvm_vcpu 121 arch/powerpc/kvm/book3s_pr.c static void kvmppc_core_vcpu_put_pr(struct kvm_vcpu *vcpu) kvm_vcpu 149 arch/powerpc/kvm/book3s_pr.c void kvmppc_copy_to_svcpu(struct kvm_vcpu *vcpu) kvm_vcpu 188 arch/powerpc/kvm/book3s_pr.c static void kvmppc_recalc_shadow_msr(struct kvm_vcpu *vcpu) kvm_vcpu 221 arch/powerpc/kvm/book3s_pr.c void kvmppc_copy_from_svcpu(struct kvm_vcpu *vcpu) kvm_vcpu 300 arch/powerpc/kvm/book3s_pr.c void kvmppc_save_tm_sprs(struct kvm_vcpu *vcpu) kvm_vcpu 309 arch/powerpc/kvm/book3s_pr.c void kvmppc_restore_tm_sprs(struct kvm_vcpu *vcpu) kvm_vcpu 321 arch/powerpc/kvm/book3s_pr.c static void kvmppc_handle_lost_math_exts(struct kvm_vcpu *vcpu) kvm_vcpu 340 arch/powerpc/kvm/book3s_pr.c void kvmppc_save_tm_pr(struct kvm_vcpu *vcpu) kvm_vcpu 355 arch/powerpc/kvm/book3s_pr.c void kvmppc_restore_tm_pr(struct kvm_vcpu *vcpu) kvm_vcpu 379 arch/powerpc/kvm/book3s_pr.c static int kvmppc_core_check_requests_pr(struct kvm_vcpu *vcpu) kvm_vcpu 396 arch/powerpc/kvm/book3s_pr.c struct kvm_vcpu *vcpu; kvm_vcpu 451 arch/powerpc/kvm/book3s_pr.c static void kvmppc_set_msr_pr(struct kvm_vcpu *vcpu, u64 msr) kvm_vcpu 536 arch/powerpc/kvm/book3s_pr.c void kvmppc_set_pvr_pr(struct kvm_vcpu *vcpu, u32 pvr) kvm_vcpu 625 arch/powerpc/kvm/book3s_pr.c static void kvmppc_patch_dcbz(struct kvm_vcpu *vcpu, struct kvmppc_pte *pte) kvm_vcpu 652 arch/powerpc/kvm/book3s_pr.c static bool kvmppc_visible_gpa(struct kvm_vcpu *vcpu, gpa_t gpa) kvm_vcpu 667 arch/powerpc/kvm/book3s_pr.c int kvmppc_handle_pagefault(struct kvm_run *run, struct kvm_vcpu *vcpu, kvm_vcpu 784 arch/powerpc/kvm/book3s_pr.c void kvmppc_giveup_ext(struct kvm_vcpu *vcpu, ulong msr) kvm_vcpu 827 arch/powerpc/kvm/book3s_pr.c void kvmppc_giveup_fac(struct kvm_vcpu *vcpu, ulong fac) kvm_vcpu 846 arch/powerpc/kvm/book3s_pr.c static int kvmppc_handle_ext(struct kvm_vcpu *vcpu, unsigned int exit_nr, kvm_vcpu 917 arch/powerpc/kvm/book3s_pr.c static void kvmppc_handle_lost_ext(struct kvm_vcpu *vcpu) kvm_vcpu 946 arch/powerpc/kvm/book3s_pr.c void kvmppc_trigger_fac_interrupt(struct kvm_vcpu *vcpu, ulong fac) kvm_vcpu 954 arch/powerpc/kvm/book3s_pr.c static void kvmppc_emulate_fac(struct kvm_vcpu *vcpu, ulong fac) kvm_vcpu 968 arch/powerpc/kvm/book3s_pr.c static int kvmppc_handle_fac(struct kvm_vcpu *vcpu, ulong fac) kvm_vcpu 1023 arch/powerpc/kvm/book3s_pr.c void kvmppc_set_fscr(struct kvm_vcpu *vcpu, u64 fscr) kvm_vcpu 1038 arch/powerpc/kvm/book3s_pr.c static void kvmppc_setup_debug(struct kvm_vcpu *vcpu) kvm_vcpu 1047 arch/powerpc/kvm/book3s_pr.c static void kvmppc_clear_debug(struct kvm_vcpu *vcpu) kvm_vcpu 1056 arch/powerpc/kvm/book3s_pr.c static int kvmppc_exit_pr_progint(struct kvm_run *run, struct kvm_vcpu *vcpu, kvm_vcpu 1119 arch/powerpc/kvm/book3s_pr.c int kvmppc_handle_exit_pr(struct kvm_run *run, struct kvm_vcpu *vcpu, kvm_vcpu 1439 arch/powerpc/kvm/book3s_pr.c static int kvm_arch_vcpu_ioctl_get_sregs_pr(struct kvm_vcpu *vcpu, kvm_vcpu 1466 arch/powerpc/kvm/book3s_pr.c static int kvm_arch_vcpu_ioctl_set_sregs_pr(struct kvm_vcpu *vcpu, kvm_vcpu 1512 arch/powerpc/kvm/book3s_pr.c static int kvmppc_get_one_reg_pr(struct kvm_vcpu *vcpu, u64 id, kvm_vcpu 1612 arch/powerpc/kvm/book3s_pr.c static void kvmppc_set_lpcr_pr(struct kvm_vcpu *vcpu, u64 new_lpcr) kvm_vcpu 1620 arch/powerpc/kvm/book3s_pr.c static int kvmppc_set_one_reg_pr(struct kvm_vcpu *vcpu, u64 id, kvm_vcpu 1711 arch/powerpc/kvm/book3s_pr.c static struct kvm_vcpu *kvmppc_core_vcpu_create_pr(struct kvm *kvm, kvm_vcpu 1715 arch/powerpc/kvm/book3s_pr.c struct kvm_vcpu *vcpu; kvm_vcpu 1792 arch/powerpc/kvm/book3s_pr.c static void kvmppc_core_vcpu_free_pr(struct kvm_vcpu *vcpu) kvm_vcpu 1805 arch/powerpc/kvm/book3s_pr.c static int kvmppc_vcpu_run_pr(struct kvm_run *kvm_run, struct kvm_vcpu *vcpu) kvm_vcpu 1867 arch/powerpc/kvm/book3s_pr.c struct kvm_vcpu *vcpu; kvm_vcpu 1940 arch/powerpc/kvm/book3s_pr.c struct kvm_vcpu *vcpu; kvm_vcpu 23 arch/powerpc/kvm/book3s_pr_papr.c static unsigned long get_pteg_addr(struct kvm_vcpu *vcpu, long pte_index) kvm_vcpu 36 arch/powerpc/kvm/book3s_pr_papr.c static int kvmppc_h_pr_enter(struct kvm_vcpu *vcpu) kvm_vcpu 86 arch/powerpc/kvm/book3s_pr_papr.c static int kvmppc_h_pr_remove(struct kvm_vcpu *vcpu) kvm_vcpu 145 arch/powerpc/kvm/book3s_pr_papr.c static int kvmppc_h_pr_bulk_remove(struct kvm_vcpu *vcpu) kvm_vcpu 214 arch/powerpc/kvm/book3s_pr_papr.c static int kvmppc_h_pr_protect(struct kvm_vcpu *vcpu) kvm_vcpu 262 arch/powerpc/kvm/book3s_pr_papr.c static int kvmppc_h_pr_logical_ci_load(struct kvm_vcpu *vcpu) kvm_vcpu 273 arch/powerpc/kvm/book3s_pr_papr.c static int kvmppc_h_pr_logical_ci_store(struct kvm_vcpu *vcpu) kvm_vcpu 285 arch/powerpc/kvm/book3s_pr_papr.c static int kvmppc_h_pr_put_tce(struct kvm_vcpu *vcpu) kvm_vcpu 299 arch/powerpc/kvm/book3s_pr_papr.c static int kvmppc_h_pr_put_tce_indirect(struct kvm_vcpu *vcpu) kvm_vcpu 315 arch/powerpc/kvm/book3s_pr_papr.c static int kvmppc_h_pr_stuff_tce(struct kvm_vcpu *vcpu) kvm_vcpu 331 arch/powerpc/kvm/book3s_pr_papr.c static int kvmppc_h_pr_put_tce(struct kvm_vcpu *vcpu) kvm_vcpu 336 arch/powerpc/kvm/book3s_pr_papr.c static int kvmppc_h_pr_put_tce_indirect(struct kvm_vcpu *vcpu) kvm_vcpu 341 arch/powerpc/kvm/book3s_pr_papr.c static int kvmppc_h_pr_stuff_tce(struct kvm_vcpu *vcpu) kvm_vcpu 347 arch/powerpc/kvm/book3s_pr_papr.c static int kvmppc_h_pr_xics_hcall(struct kvm_vcpu *vcpu, u32 cmd) kvm_vcpu 354 arch/powerpc/kvm/book3s_pr_papr.c int kvmppc_h_pr(struct kvm_vcpu *vcpu, unsigned long cmd) kvm_vcpu 19 arch/powerpc/kvm/book3s_rtas.c static void kvm_rtas_set_xive(struct kvm_vcpu *vcpu, struct rtas_args *args) kvm_vcpu 43 arch/powerpc/kvm/book3s_rtas.c static void kvm_rtas_get_xive(struct kvm_vcpu *vcpu, struct rtas_args *args) kvm_vcpu 71 arch/powerpc/kvm/book3s_rtas.c static void kvm_rtas_int_off(struct kvm_vcpu *vcpu, struct rtas_args *args) kvm_vcpu 93 arch/powerpc/kvm/book3s_rtas.c static void kvm_rtas_int_on(struct kvm_vcpu *vcpu, struct rtas_args *args) kvm_vcpu 117 arch/powerpc/kvm/book3s_rtas.c void (*handler)(struct kvm_vcpu *vcpu, struct rtas_args *args); kvm_vcpu 218 arch/powerpc/kvm/book3s_rtas.c int kvmppc_rtas_hcall(struct kvm_vcpu *vcpu) kvm_vcpu 586 arch/powerpc/kvm/book3s_xics.c static noinline unsigned long kvmppc_h_xirr(struct kvm_vcpu *vcpu) kvm_vcpu 619 arch/powerpc/kvm/book3s_xics.c static noinline int kvmppc_h_ipi(struct kvm_vcpu *vcpu, unsigned long server, kvm_vcpu 704 arch/powerpc/kvm/book3s_xics.c static int kvmppc_h_ipoll(struct kvm_vcpu *vcpu, unsigned long server) kvm_vcpu 721 arch/powerpc/kvm/book3s_xics.c static noinline void kvmppc_h_cppr(struct kvm_vcpu *vcpu, unsigned long cppr) kvm_vcpu 777 arch/powerpc/kvm/book3s_xics.c static int ics_eoi(struct kvm_vcpu *vcpu, u32 irq) kvm_vcpu 817 arch/powerpc/kvm/book3s_xics.c static noinline int kvmppc_h_eoi(struct kvm_vcpu *vcpu, unsigned long xirr) kvm_vcpu 848 arch/powerpc/kvm/book3s_xics.c int kvmppc_xics_rm_complete(struct kvm_vcpu *vcpu, u32 hcall) kvm_vcpu 875 arch/powerpc/kvm/book3s_xics.c int kvmppc_xics_hcall(struct kvm_vcpu *vcpu, u32 req) kvm_vcpu 944 arch/powerpc/kvm/book3s_xics.c struct kvm_vcpu *vcpu; kvm_vcpu 1071 arch/powerpc/kvm/book3s_xics.c static int kvmppc_xics_create_icp(struct kvm_vcpu *vcpu, unsigned long server_num) kvm_vcpu 1096 arch/powerpc/kvm/book3s_xics.c u64 kvmppc_xics_get_icp(struct kvm_vcpu *vcpu) kvm_vcpu 1110 arch/powerpc/kvm/book3s_xics.c int kvmppc_xics_set_icp(struct kvm_vcpu *vcpu, u64 icpval) kvm_vcpu 1408 arch/powerpc/kvm/book3s_xics.c int kvmppc_xics_connect_vcpu(struct kvm_device *dev, struct kvm_vcpu *vcpu, kvm_vcpu 1428 arch/powerpc/kvm/book3s_xics.c void kvmppc_xics_free_icp(struct kvm_vcpu *vcpu) kvm_vcpu 67 arch/powerpc/kvm/book3s_xics.h struct kvm_vcpu *vcpu; kvm_vcpu 79 arch/powerpc/kvm/book3s_xics.h struct kvm_vcpu *rm_kick_target; kvm_vcpu 94 arch/powerpc/kvm/book3s_xics.h struct kvm_vcpu *rm_dbgtgt; kvm_vcpu 118 arch/powerpc/kvm/book3s_xics.h struct kvm_vcpu *vcpu = NULL; kvm_vcpu 145 arch/powerpc/kvm/book3s_xics.h extern unsigned long xics_rm_h_xirr(struct kvm_vcpu *vcpu); kvm_vcpu 146 arch/powerpc/kvm/book3s_xics.h extern int xics_rm_h_ipi(struct kvm_vcpu *vcpu, unsigned long server, kvm_vcpu 148 arch/powerpc/kvm/book3s_xics.h extern int xics_rm_h_cppr(struct kvm_vcpu *vcpu, unsigned long cppr); kvm_vcpu 149 arch/powerpc/kvm/book3s_xics.h extern int xics_rm_h_eoi(struct kvm_vcpu *vcpu, unsigned long xirr); kvm_vcpu 65 arch/powerpc/kvm/book3s_xive.c void kvmppc_xive_push_vcpu(struct kvm_vcpu *vcpu) kvm_vcpu 151 arch/powerpc/kvm/book3s_xive.c struct kvm_vcpu *vcpu = data; kvm_vcpu 175 arch/powerpc/kvm/book3s_xive.c int kvmppc_xive_attach_escalation(struct kvm_vcpu *vcpu, u8 prio, kvm_vcpu 245 arch/powerpc/kvm/book3s_xive.c static int xive_provision_queue(struct kvm_vcpu *vcpu, u8 prio) kvm_vcpu 284 arch/powerpc/kvm/book3s_xive.c struct kvm_vcpu *vcpu; kvm_vcpu 315 arch/powerpc/kvm/book3s_xive.c struct kvm_vcpu *vcpu; kvm_vcpu 333 arch/powerpc/kvm/book3s_xive.c static int xive_try_pick_queue(struct kvm_vcpu *vcpu, u8 prio) kvm_vcpu 355 arch/powerpc/kvm/book3s_xive.c struct kvm_vcpu *vcpu; kvm_vcpu 824 arch/powerpc/kvm/book3s_xive.c u64 kvmppc_xive_get_icp(struct kvm_vcpu *vcpu) kvm_vcpu 837 arch/powerpc/kvm/book3s_xive.c int kvmppc_xive_set_icp(struct kvm_vcpu *vcpu, u64 icpval) kvm_vcpu 1074 arch/powerpc/kvm/book3s_xive.c void kvmppc_xive_disable_vcpu_interrupts(struct kvm_vcpu *vcpu) kvm_vcpu 1133 arch/powerpc/kvm/book3s_xive.c void xive_cleanup_single_escalation(struct kvm_vcpu *vcpu, kvm_vcpu 1150 arch/powerpc/kvm/book3s_xive.c void kvmppc_xive_cleanup_vcpu(struct kvm_vcpu *vcpu) kvm_vcpu 1215 arch/powerpc/kvm/book3s_xive.c struct kvm_vcpu *vcpu, u32 cpu) kvm_vcpu 1438 arch/powerpc/kvm/book3s_xive.c struct kvm_vcpu *vcpu = NULL; kvm_vcpu 1619 arch/powerpc/kvm/book3s_xive.c struct kvm_vcpu *vcpu = NULL; kvm_vcpu 1909 arch/powerpc/kvm/book3s_xive.c struct kvm_vcpu *vcpu; kvm_vcpu 2044 arch/powerpc/kvm/book3s_xive.c int kvmppc_xive_debug_show_queues(struct seq_file *m, struct kvm_vcpu *vcpu) kvm_vcpu 2086 arch/powerpc/kvm/book3s_xive.c struct kvm_vcpu *vcpu; kvm_vcpu 148 arch/powerpc/kvm/book3s_xive.h struct kvm_vcpu *vcpu; kvm_vcpu 193 arch/powerpc/kvm/book3s_xive.h static inline struct kvm_vcpu *kvmppc_xive_find_server(struct kvm *kvm, u32 nr) kvm_vcpu 195 arch/powerpc/kvm/book3s_xive.h struct kvm_vcpu *vcpu = NULL; kvm_vcpu 225 arch/powerpc/kvm/book3s_xive.h struct kvm_vcpu *vcpu = NULL; kvm_vcpu 271 arch/powerpc/kvm/book3s_xive.h extern unsigned long xive_rm_h_xirr(struct kvm_vcpu *vcpu); kvm_vcpu 272 arch/powerpc/kvm/book3s_xive.h extern unsigned long xive_rm_h_ipoll(struct kvm_vcpu *vcpu, unsigned long server); kvm_vcpu 273 arch/powerpc/kvm/book3s_xive.h extern int xive_rm_h_ipi(struct kvm_vcpu *vcpu, unsigned long server, kvm_vcpu 275 arch/powerpc/kvm/book3s_xive.h extern int xive_rm_h_cppr(struct kvm_vcpu *vcpu, unsigned long cppr); kvm_vcpu 276 arch/powerpc/kvm/book3s_xive.h extern int xive_rm_h_eoi(struct kvm_vcpu *vcpu, unsigned long xirr); kvm_vcpu 278 arch/powerpc/kvm/book3s_xive.h extern unsigned long (*__xive_vm_h_xirr)(struct kvm_vcpu *vcpu); kvm_vcpu 279 arch/powerpc/kvm/book3s_xive.h extern unsigned long (*__xive_vm_h_ipoll)(struct kvm_vcpu *vcpu, unsigned long server); kvm_vcpu 280 arch/powerpc/kvm/book3s_xive.h extern int (*__xive_vm_h_ipi)(struct kvm_vcpu *vcpu, unsigned long server, kvm_vcpu 282 arch/powerpc/kvm/book3s_xive.h extern int (*__xive_vm_h_cppr)(struct kvm_vcpu *vcpu, unsigned long cppr); kvm_vcpu 283 arch/powerpc/kvm/book3s_xive.h extern int (*__xive_vm_h_eoi)(struct kvm_vcpu *vcpu, unsigned long xirr); kvm_vcpu 288 arch/powerpc/kvm/book3s_xive.h void kvmppc_xive_disable_vcpu_interrupts(struct kvm_vcpu *vcpu); kvm_vcpu 289 arch/powerpc/kvm/book3s_xive.h int kvmppc_xive_debug_show_queues(struct seq_file *m, struct kvm_vcpu *vcpu); kvm_vcpu 294 arch/powerpc/kvm/book3s_xive.h int kvmppc_xive_attach_escalation(struct kvm_vcpu *vcpu, u8 prio, kvm_vcpu 297 arch/powerpc/kvm/book3s_xive.h void xive_cleanup_single_escalation(struct kvm_vcpu *vcpu, kvm_vcpu 41 arch/powerpc/kvm/book3s_xive_native.c static void kvmppc_xive_native_cleanup_queue(struct kvm_vcpu *vcpu, int prio) kvm_vcpu 71 arch/powerpc/kvm/book3s_xive_native.c void kvmppc_xive_native_cleanup_vcpu(struct kvm_vcpu *vcpu) kvm_vcpu 122 arch/powerpc/kvm/book3s_xive_native.c struct kvm_vcpu *vcpu, u32 server_num) kvm_vcpu 555 arch/powerpc/kvm/book3s_xive_native.c struct kvm_vcpu *vcpu; kvm_vcpu 703 arch/powerpc/kvm/book3s_xive_native.c struct kvm_vcpu *vcpu; kvm_vcpu 800 arch/powerpc/kvm/book3s_xive_native.c struct kvm_vcpu *vcpu; kvm_vcpu 883 arch/powerpc/kvm/book3s_xive_native.c static int kvmppc_xive_native_vcpu_eq_sync(struct kvm_vcpu *vcpu) kvm_vcpu 909 arch/powerpc/kvm/book3s_xive_native.c struct kvm_vcpu *vcpu; kvm_vcpu 1007 arch/powerpc/kvm/book3s_xive_native.c struct kvm_vcpu *vcpu; kvm_vcpu 1128 arch/powerpc/kvm/book3s_xive_native.c int kvmppc_xive_native_get_vp(struct kvm_vcpu *vcpu, union kvmppc_one_reg *val) kvm_vcpu 1166 arch/powerpc/kvm/book3s_xive_native.c int kvmppc_xive_native_set_vp(struct kvm_vcpu *vcpu, union kvmppc_one_reg *val) kvm_vcpu 1207 arch/powerpc/kvm/book3s_xive_native.c struct kvm_vcpu *vcpu; kvm_vcpu 269 arch/powerpc/kvm/book3s_xive_template.c X_STATIC unsigned long GLUE(X_PFX,h_xirr)(struct kvm_vcpu *vcpu) kvm_vcpu 320 arch/powerpc/kvm/book3s_xive_template.c X_STATIC unsigned long GLUE(X_PFX,h_ipoll)(struct kvm_vcpu *vcpu, unsigned long server) kvm_vcpu 439 arch/powerpc/kvm/book3s_xive_template.c X_STATIC int GLUE(X_PFX,h_cppr)(struct kvm_vcpu *vcpu, unsigned long cppr) kvm_vcpu 498 arch/powerpc/kvm/book3s_xive_template.c X_STATIC int GLUE(X_PFX,h_eoi)(struct kvm_vcpu *vcpu, unsigned long xirr) kvm_vcpu 603 arch/powerpc/kvm/book3s_xive_template.c X_STATIC int GLUE(X_PFX,h_ipi)(struct kvm_vcpu *vcpu, unsigned long server, kvm_vcpu 39 arch/powerpc/kvm/booke.c #define VCPU_STAT(x) offsetof(struct kvm_vcpu, stat.x), KVM_STAT_VCPU kvm_vcpu 65 arch/powerpc/kvm/booke.c void kvmppc_dump_vcpu(struct kvm_vcpu *vcpu) kvm_vcpu 88 arch/powerpc/kvm/booke.c void kvmppc_vcpu_disable_spe(struct kvm_vcpu *vcpu) kvm_vcpu 98 arch/powerpc/kvm/booke.c static void kvmppc_vcpu_enable_spe(struct kvm_vcpu *vcpu) kvm_vcpu 108 arch/powerpc/kvm/booke.c static void kvmppc_vcpu_sync_spe(struct kvm_vcpu *vcpu) kvm_vcpu 118 arch/powerpc/kvm/booke.c static void kvmppc_vcpu_sync_spe(struct kvm_vcpu *vcpu) kvm_vcpu 132 arch/powerpc/kvm/booke.c static inline void kvmppc_load_guest_fp(struct kvm_vcpu *vcpu) kvm_vcpu 149 arch/powerpc/kvm/booke.c static inline void kvmppc_save_guest_fp(struct kvm_vcpu *vcpu) kvm_vcpu 158 arch/powerpc/kvm/booke.c static void kvmppc_vcpu_sync_fpu(struct kvm_vcpu *vcpu) kvm_vcpu 173 arch/powerpc/kvm/booke.c static inline void kvmppc_load_guest_altivec(struct kvm_vcpu *vcpu) kvm_vcpu 192 arch/powerpc/kvm/booke.c static inline void kvmppc_save_guest_altivec(struct kvm_vcpu *vcpu) kvm_vcpu 203 arch/powerpc/kvm/booke.c static void kvmppc_vcpu_sync_debug(struct kvm_vcpu *vcpu) kvm_vcpu 230 arch/powerpc/kvm/booke.c void kvmppc_set_msr(struct kvm_vcpu *vcpu, u32 new_msr) kvm_vcpu 246 arch/powerpc/kvm/booke.c static void kvmppc_booke_queue_irqprio(struct kvm_vcpu *vcpu, kvm_vcpu 253 arch/powerpc/kvm/booke.c void kvmppc_core_queue_dtlb_miss(struct kvm_vcpu *vcpu, kvm_vcpu 261 arch/powerpc/kvm/booke.c void kvmppc_core_queue_data_storage(struct kvm_vcpu *vcpu, kvm_vcpu 269 arch/powerpc/kvm/booke.c void kvmppc_core_queue_itlb_miss(struct kvm_vcpu *vcpu) kvm_vcpu 274 arch/powerpc/kvm/booke.c void kvmppc_core_queue_inst_storage(struct kvm_vcpu *vcpu, ulong esr_flags) kvm_vcpu 280 arch/powerpc/kvm/booke.c static void kvmppc_core_queue_alignment(struct kvm_vcpu *vcpu, ulong dear_flags, kvm_vcpu 288 arch/powerpc/kvm/booke.c void kvmppc_core_queue_program(struct kvm_vcpu *vcpu, ulong esr_flags) kvm_vcpu 294 arch/powerpc/kvm/booke.c void kvmppc_core_queue_fpunavail(struct kvm_vcpu *vcpu) kvm_vcpu 300 arch/powerpc/kvm/booke.c void kvmppc_core_queue_vec_unavail(struct kvm_vcpu *vcpu) kvm_vcpu 306 arch/powerpc/kvm/booke.c void kvmppc_core_queue_dec(struct kvm_vcpu *vcpu) kvm_vcpu 311 arch/powerpc/kvm/booke.c int kvmppc_core_pending_dec(struct kvm_vcpu *vcpu) kvm_vcpu 316 arch/powerpc/kvm/booke.c void kvmppc_core_dequeue_dec(struct kvm_vcpu *vcpu) kvm_vcpu 321 arch/powerpc/kvm/booke.c void kvmppc_core_queue_external(struct kvm_vcpu *vcpu, kvm_vcpu 332 arch/powerpc/kvm/booke.c void kvmppc_core_dequeue_external(struct kvm_vcpu *vcpu) kvm_vcpu 338 arch/powerpc/kvm/booke.c static void kvmppc_core_queue_watchdog(struct kvm_vcpu *vcpu) kvm_vcpu 343 arch/powerpc/kvm/booke.c static void kvmppc_core_dequeue_watchdog(struct kvm_vcpu *vcpu) kvm_vcpu 348 arch/powerpc/kvm/booke.c void kvmppc_core_queue_debug(struct kvm_vcpu *vcpu) kvm_vcpu 353 arch/powerpc/kvm/booke.c void kvmppc_core_dequeue_debug(struct kvm_vcpu *vcpu) kvm_vcpu 358 arch/powerpc/kvm/booke.c static void set_guest_srr(struct kvm_vcpu *vcpu, unsigned long srr0, u32 srr1) kvm_vcpu 364 arch/powerpc/kvm/booke.c static void set_guest_csrr(struct kvm_vcpu *vcpu, unsigned long srr0, u32 srr1) kvm_vcpu 370 arch/powerpc/kvm/booke.c static void set_guest_dsrr(struct kvm_vcpu *vcpu, unsigned long srr0, u32 srr1) kvm_vcpu 380 arch/powerpc/kvm/booke.c static void set_guest_mcsrr(struct kvm_vcpu *vcpu, unsigned long srr0, u32 srr1) kvm_vcpu 387 arch/powerpc/kvm/booke.c static int kvmppc_booke_irqprio_deliver(struct kvm_vcpu *vcpu, kvm_vcpu 550 arch/powerpc/kvm/booke.c static unsigned long watchdog_next_timeout(struct kvm_vcpu *vcpu) kvm_vcpu 576 arch/powerpc/kvm/booke.c static void arm_next_watchdog(struct kvm_vcpu *vcpu) kvm_vcpu 603 arch/powerpc/kvm/booke.c struct kvm_vcpu *vcpu = from_timer(vcpu, t, arch.wdt_timer); kvm_vcpu 649 arch/powerpc/kvm/booke.c static void update_timer_ints(struct kvm_vcpu *vcpu) kvm_vcpu 662 arch/powerpc/kvm/booke.c static void kvmppc_core_check_exceptions(struct kvm_vcpu *vcpu) kvm_vcpu 682 arch/powerpc/kvm/booke.c int kvmppc_core_prepare_to_enter(struct kvm_vcpu *vcpu) kvm_vcpu 707 arch/powerpc/kvm/booke.c int kvmppc_core_check_requests(struct kvm_vcpu *vcpu) kvm_vcpu 733 arch/powerpc/kvm/booke.c int kvmppc_vcpu_run(struct kvm_run *kvm_run, struct kvm_vcpu *vcpu) kvm_vcpu 803 arch/powerpc/kvm/booke.c static int emulation_exit(struct kvm_run *run, struct kvm_vcpu *vcpu) kvm_vcpu 837 arch/powerpc/kvm/booke.c static int kvmppc_handle_debug(struct kvm_run *run, struct kvm_vcpu *vcpu) kvm_vcpu 910 arch/powerpc/kvm/booke.c static void kvmppc_restart_interrupt(struct kvm_vcpu *vcpu, kvm_vcpu 957 arch/powerpc/kvm/booke.c static int kvmppc_resume_inst_load(struct kvm_run *run, struct kvm_vcpu *vcpu, kvm_vcpu 984 arch/powerpc/kvm/booke.c int kvmppc_handle_exit(struct kvm_run *run, struct kvm_vcpu *vcpu, kvm_vcpu 1368 arch/powerpc/kvm/booke.c static void kvmppc_set_tsr(struct kvm_vcpu *vcpu, u32 new_tsr) kvm_vcpu 1381 arch/powerpc/kvm/booke.c int kvm_arch_vcpu_setup(struct kvm_vcpu *vcpu) kvm_vcpu 1410 arch/powerpc/kvm/booke.c int kvmppc_subarch_vcpu_init(struct kvm_vcpu *vcpu) kvm_vcpu 1424 arch/powerpc/kvm/booke.c void kvmppc_subarch_vcpu_uninit(struct kvm_vcpu *vcpu) kvm_vcpu 1429 arch/powerpc/kvm/booke.c int kvm_arch_vcpu_ioctl_get_regs(struct kvm_vcpu *vcpu, struct kvm_regs *regs) kvm_vcpu 1460 arch/powerpc/kvm/booke.c int kvm_arch_vcpu_ioctl_set_regs(struct kvm_vcpu *vcpu, struct kvm_regs *regs) kvm_vcpu 1491 arch/powerpc/kvm/booke.c static void get_sregs_base(struct kvm_vcpu *vcpu, kvm_vcpu 1510 arch/powerpc/kvm/booke.c static int set_sregs_base(struct kvm_vcpu *vcpu, kvm_vcpu 1535 arch/powerpc/kvm/booke.c static void get_sregs_arch206(struct kvm_vcpu *vcpu, kvm_vcpu 1547 arch/powerpc/kvm/booke.c static int set_sregs_arch206(struct kvm_vcpu *vcpu, kvm_vcpu 1564 arch/powerpc/kvm/booke.c int kvmppc_get_sregs_ivor(struct kvm_vcpu *vcpu, struct kvm_sregs *sregs) kvm_vcpu 1587 arch/powerpc/kvm/booke.c int kvmppc_set_sregs_ivor(struct kvm_vcpu *vcpu, struct kvm_sregs *sregs) kvm_vcpu 1612 arch/powerpc/kvm/booke.c int kvm_arch_vcpu_ioctl_get_sregs(struct kvm_vcpu *vcpu, kvm_vcpu 1629 arch/powerpc/kvm/booke.c int kvm_arch_vcpu_ioctl_set_sregs(struct kvm_vcpu *vcpu, kvm_vcpu 1653 arch/powerpc/kvm/booke.c int kvmppc_get_one_reg(struct kvm_vcpu *vcpu, u64 id, kvm_vcpu 1709 arch/powerpc/kvm/booke.c int kvmppc_set_one_reg(struct kvm_vcpu *vcpu, u64 id, kvm_vcpu 1778 arch/powerpc/kvm/booke.c int kvm_arch_vcpu_ioctl_get_fpu(struct kvm_vcpu *vcpu, struct kvm_fpu *fpu) kvm_vcpu 1783 arch/powerpc/kvm/booke.c int kvm_arch_vcpu_ioctl_set_fpu(struct kvm_vcpu *vcpu, struct kvm_fpu *fpu) kvm_vcpu 1788 arch/powerpc/kvm/booke.c int kvm_arch_vcpu_ioctl_translate(struct kvm_vcpu *vcpu, kvm_vcpu 1834 arch/powerpc/kvm/booke.c void kvmppc_set_epcr(struct kvm_vcpu *vcpu, u32 new_epcr) kvm_vcpu 1846 arch/powerpc/kvm/booke.c void kvmppc_set_tcr(struct kvm_vcpu *vcpu, u32 new_tcr) kvm_vcpu 1853 arch/powerpc/kvm/booke.c void kvmppc_set_tsr_bits(struct kvm_vcpu *vcpu, u32 tsr_bits) kvm_vcpu 1861 arch/powerpc/kvm/booke.c void kvmppc_clr_tsr_bits(struct kvm_vcpu *vcpu, u32 tsr_bits) kvm_vcpu 1875 arch/powerpc/kvm/booke.c void kvmppc_decrementer_func(struct kvm_vcpu *vcpu) kvm_vcpu 1940 arch/powerpc/kvm/booke.c void kvm_guest_protect_msr(struct kvm_vcpu *vcpu, ulong prot_bitmap, bool set) kvm_vcpu 1963 arch/powerpc/kvm/booke.c int kvmppc_xlate(struct kvm_vcpu *vcpu, ulong eaddr, enum xlate_instdata xlid, kvm_vcpu 2014 arch/powerpc/kvm/booke.c int kvm_arch_vcpu_ioctl_set_guest_debug(struct kvm_vcpu *vcpu, kvm_vcpu 2092 arch/powerpc/kvm/booke.c void kvmppc_booke_vcpu_load(struct kvm_vcpu *vcpu, int cpu) kvm_vcpu 2095 arch/powerpc/kvm/booke.c current->thread.kvm_vcpu = vcpu; kvm_vcpu 2098 arch/powerpc/kvm/booke.c void kvmppc_booke_vcpu_put(struct kvm_vcpu *vcpu) kvm_vcpu 2100 arch/powerpc/kvm/booke.c current->thread.kvm_vcpu = NULL; kvm_vcpu 2107 arch/powerpc/kvm/booke.c void kvmppc_mmu_destroy(struct kvm_vcpu *vcpu) kvm_vcpu 2117 arch/powerpc/kvm/booke.c struct kvm_vcpu *kvmppc_core_vcpu_create(struct kvm *kvm, unsigned int id) kvm_vcpu 2122 arch/powerpc/kvm/booke.c void kvmppc_core_vcpu_free(struct kvm_vcpu *vcpu) kvm_vcpu 2132 arch/powerpc/kvm/booke.c void kvmppc_core_vcpu_load(struct kvm_vcpu *vcpu, int cpu) kvm_vcpu 2137 arch/powerpc/kvm/booke.c void kvmppc_core_vcpu_put(struct kvm_vcpu *vcpu) kvm_vcpu 65 arch/powerpc/kvm/booke.h void kvmppc_set_msr(struct kvm_vcpu *vcpu, u32 new_msr); kvm_vcpu 66 arch/powerpc/kvm/booke.h void kvmppc_mmu_msr_notify(struct kvm_vcpu *vcpu, u32 old_msr); kvm_vcpu 68 arch/powerpc/kvm/booke.h void kvmppc_set_epcr(struct kvm_vcpu *vcpu, u32 new_epcr); kvm_vcpu 69 arch/powerpc/kvm/booke.h void kvmppc_set_tcr(struct kvm_vcpu *vcpu, u32 new_tcr); kvm_vcpu 70 arch/powerpc/kvm/booke.h void kvmppc_set_tsr_bits(struct kvm_vcpu *vcpu, u32 tsr_bits); kvm_vcpu 71 arch/powerpc/kvm/booke.h void kvmppc_clr_tsr_bits(struct kvm_vcpu *vcpu, u32 tsr_bits); kvm_vcpu 73 arch/powerpc/kvm/booke.h int kvmppc_booke_emulate_op(struct kvm_run *run, struct kvm_vcpu *vcpu, kvm_vcpu 75 arch/powerpc/kvm/booke.h int kvmppc_booke_emulate_mfspr(struct kvm_vcpu *vcpu, int sprn, ulong *spr_val); kvm_vcpu 76 arch/powerpc/kvm/booke.h int kvmppc_booke_emulate_mtspr(struct kvm_vcpu *vcpu, int sprn, ulong spr_val); kvm_vcpu 79 arch/powerpc/kvm/booke.h void kvmppc_load_guest_spe(struct kvm_vcpu *vcpu); kvm_vcpu 80 arch/powerpc/kvm/booke.h void kvmppc_save_guest_spe(struct kvm_vcpu *vcpu); kvm_vcpu 83 arch/powerpc/kvm/booke.h void kvmppc_vcpu_disable_spe(struct kvm_vcpu *vcpu); kvm_vcpu 85 arch/powerpc/kvm/booke.h void kvmppc_booke_vcpu_load(struct kvm_vcpu *vcpu, int cpu); kvm_vcpu 86 arch/powerpc/kvm/booke.h void kvmppc_booke_vcpu_put(struct kvm_vcpu *vcpu); kvm_vcpu 95 arch/powerpc/kvm/booke.h void kvmppc_set_pending_interrupt(struct kvm_vcpu *vcpu, enum int_class type); kvm_vcpu 97 arch/powerpc/kvm/booke.h extern void kvmppc_mmu_destroy_e500(struct kvm_vcpu *vcpu); kvm_vcpu 99 arch/powerpc/kvm/booke.h struct kvm_vcpu *vcpu, kvm_vcpu 101 arch/powerpc/kvm/booke.h extern int kvmppc_core_emulate_mtspr_e500(struct kvm_vcpu *vcpu, int sprn, kvm_vcpu 103 arch/powerpc/kvm/booke.h extern int kvmppc_core_emulate_mfspr_e500(struct kvm_vcpu *vcpu, int sprn, kvm_vcpu 105 arch/powerpc/kvm/booke.h extern void kvmppc_mmu_destroy_e500(struct kvm_vcpu *vcpu); kvm_vcpu 107 arch/powerpc/kvm/booke.h struct kvm_vcpu *vcpu, kvm_vcpu 109 arch/powerpc/kvm/booke.h extern int kvmppc_core_emulate_mtspr_e500(struct kvm_vcpu *vcpu, int sprn, kvm_vcpu 111 arch/powerpc/kvm/booke.h extern int kvmppc_core_emulate_mfspr_e500(struct kvm_vcpu *vcpu, int sprn, kvm_vcpu 24 arch/powerpc/kvm/booke_emulate.c static void kvmppc_emul_rfi(struct kvm_vcpu *vcpu) kvm_vcpu 30 arch/powerpc/kvm/booke_emulate.c static void kvmppc_emul_rfdi(struct kvm_vcpu *vcpu) kvm_vcpu 36 arch/powerpc/kvm/booke_emulate.c static void kvmppc_emul_rfci(struct kvm_vcpu *vcpu) kvm_vcpu 42 arch/powerpc/kvm/booke_emulate.c int kvmppc_booke_emulate_op(struct kvm_run *run, struct kvm_vcpu *vcpu, kvm_vcpu 120 arch/powerpc/kvm/booke_emulate.c int kvmppc_booke_emulate_mtspr(struct kvm_vcpu *vcpu, int sprn, ulong spr_val) kvm_vcpu 379 arch/powerpc/kvm/booke_emulate.c int kvmppc_booke_emulate_mfspr(struct kvm_vcpu *vcpu, int sprn, ulong *spr_val) kvm_vcpu 214 arch/powerpc/kvm/e500.c unsigned int kvmppc_e500_get_tlb_stid(struct kvm_vcpu *vcpu, kvm_vcpu 221 arch/powerpc/kvm/e500.c void kvmppc_set_pid(struct kvm_vcpu *vcpu, u32 pid) kvm_vcpu 293 arch/powerpc/kvm/e500.c void kvmppc_mmu_msr_notify(struct kvm_vcpu *vcpu, u32 old_msr) kvm_vcpu 299 arch/powerpc/kvm/e500.c static void kvmppc_core_vcpu_load_e500(struct kvm_vcpu *vcpu, int cpu) kvm_vcpu 307 arch/powerpc/kvm/e500.c static void kvmppc_core_vcpu_put_e500(struct kvm_vcpu *vcpu) kvm_vcpu 346 arch/powerpc/kvm/e500.c int kvmppc_core_vcpu_setup(struct kvm_vcpu *vcpu) kvm_vcpu 361 arch/powerpc/kvm/e500.c static int kvmppc_core_get_sregs_e500(struct kvm_vcpu *vcpu, kvm_vcpu 386 arch/powerpc/kvm/e500.c static int kvmppc_core_set_sregs_e500(struct kvm_vcpu *vcpu, kvm_vcpu 422 arch/powerpc/kvm/e500.c static int kvmppc_get_one_reg_e500(struct kvm_vcpu *vcpu, u64 id, kvm_vcpu 429 arch/powerpc/kvm/e500.c static int kvmppc_set_one_reg_e500(struct kvm_vcpu *vcpu, u64 id, kvm_vcpu 436 arch/powerpc/kvm/e500.c static struct kvm_vcpu *kvmppc_core_vcpu_create_e500(struct kvm *kvm, kvm_vcpu 440 arch/powerpc/kvm/e500.c struct kvm_vcpu *vcpu; kvm_vcpu 486 arch/powerpc/kvm/e500.c static void kvmppc_core_vcpu_free_e500(struct kvm_vcpu *vcpu) kvm_vcpu 58 arch/powerpc/kvm/e500.h struct kvm_vcpu vcpu; kvm_vcpu 100 arch/powerpc/kvm/e500.h static inline struct kvmppc_vcpu_e500 *to_e500(struct kvm_vcpu *vcpu) kvm_vcpu 127 arch/powerpc/kvm/e500.h int kvmppc_e500_emul_tlbwe(struct kvm_vcpu *vcpu); kvm_vcpu 128 arch/powerpc/kvm/e500.h int kvmppc_e500_emul_tlbre(struct kvm_vcpu *vcpu); kvm_vcpu 129 arch/powerpc/kvm/e500.h int kvmppc_e500_emul_tlbivax(struct kvm_vcpu *vcpu, gva_t ea); kvm_vcpu 130 arch/powerpc/kvm/e500.h int kvmppc_e500_emul_tlbilx(struct kvm_vcpu *vcpu, int type, gva_t ea); kvm_vcpu 131 arch/powerpc/kvm/e500.h int kvmppc_e500_emul_tlbsx(struct kvm_vcpu *vcpu, gva_t ea); kvm_vcpu 135 arch/powerpc/kvm/e500.h void kvmppc_get_sregs_e500_tlb(struct kvm_vcpu *vcpu, struct kvm_sregs *sregs); kvm_vcpu 136 arch/powerpc/kvm/e500.h int kvmppc_set_sregs_e500_tlb(struct kvm_vcpu *vcpu, struct kvm_sregs *sregs); kvm_vcpu 138 arch/powerpc/kvm/e500.h int kvmppc_get_one_reg_e500_tlb(struct kvm_vcpu *vcpu, u64 id, kvm_vcpu 140 arch/powerpc/kvm/e500.h int kvmppc_set_one_reg_e500_tlb(struct kvm_vcpu *vcpu, u64 id, kvm_vcpu 208 arch/powerpc/kvm/e500.h static inline unsigned int get_cur_pid(struct kvm_vcpu *vcpu) kvm_vcpu 213 arch/powerpc/kvm/e500.h static inline unsigned int get_cur_as(struct kvm_vcpu *vcpu) kvm_vcpu 218 arch/powerpc/kvm/e500.h static inline unsigned int get_cur_pr(struct kvm_vcpu *vcpu) kvm_vcpu 223 arch/powerpc/kvm/e500.h static inline unsigned int get_cur_spid(const struct kvm_vcpu *vcpu) kvm_vcpu 228 arch/powerpc/kvm/e500.h static inline unsigned int get_cur_sas(const struct kvm_vcpu *vcpu) kvm_vcpu 233 arch/powerpc/kvm/e500.h static inline unsigned int get_tlb_tlbsel(const struct kvm_vcpu *vcpu) kvm_vcpu 242 arch/powerpc/kvm/e500.h static inline unsigned int get_tlb_nv_bit(const struct kvm_vcpu *vcpu) kvm_vcpu 247 arch/powerpc/kvm/e500.h static inline unsigned int get_tlb_esel_bit(const struct kvm_vcpu *vcpu) kvm_vcpu 252 arch/powerpc/kvm/e500.h static inline int tlbe_is_host_safe(const struct kvm_vcpu *vcpu, kvm_vcpu 305 arch/powerpc/kvm/e500.h static inline int get_lpid(struct kvm_vcpu *vcpu) kvm_vcpu 310 arch/powerpc/kvm/e500.h unsigned int kvmppc_e500_get_tlb_stid(struct kvm_vcpu *vcpu, kvm_vcpu 313 arch/powerpc/kvm/e500.h static inline unsigned int get_tlbmiss_tid(struct kvm_vcpu *vcpu) kvm_vcpu 325 arch/powerpc/kvm/e500.h static inline bool has_feature(const struct kvm_vcpu *vcpu, kvm_vcpu 51 arch/powerpc/kvm/e500_emulate.c static int kvmppc_e500_emul_msgclr(struct kvm_vcpu *vcpu, int rb) kvm_vcpu 63 arch/powerpc/kvm/e500_emulate.c static int kvmppc_e500_emul_msgsnd(struct kvm_vcpu *vcpu, int rb) kvm_vcpu 69 arch/powerpc/kvm/e500_emulate.c struct kvm_vcpu *cvcpu; kvm_vcpu 86 arch/powerpc/kvm/e500_emulate.c static int kvmppc_e500_emul_ehpriv(struct kvm_run *run, struct kvm_vcpu *vcpu, kvm_vcpu 106 arch/powerpc/kvm/e500_emulate.c static int kvmppc_e500_emul_dcbtls(struct kvm_vcpu *vcpu) kvm_vcpu 115 arch/powerpc/kvm/e500_emulate.c static int kvmppc_e500_emul_mftmr(struct kvm_vcpu *vcpu, unsigned int inst, kvm_vcpu 128 arch/powerpc/kvm/e500_emulate.c int kvmppc_core_emulate_op_e500(struct kvm_run *run, struct kvm_vcpu *vcpu, kvm_vcpu 205 arch/powerpc/kvm/e500_emulate.c int kvmppc_core_emulate_mtspr_e500(struct kvm_vcpu *vcpu, int sprn, ulong spr_val) kvm_vcpu 322 arch/powerpc/kvm/e500_emulate.c int kvmppc_core_emulate_mfspr_e500(struct kvm_vcpu *vcpu, int sprn, ulong *spr_val) kvm_vcpu 64 arch/powerpc/kvm/e500_mmu.c static unsigned int get_tlb_esel(struct kvm_vcpu *vcpu, int tlbsel) kvm_vcpu 126 arch/powerpc/kvm/e500_mmu.c static inline void kvmppc_e500_deliver_tlb_miss(struct kvm_vcpu *vcpu, kvm_vcpu 193 arch/powerpc/kvm/e500_mmu.c static void kvmppc_set_tlb1map_range(struct kvm_vcpu *vcpu, kvm_vcpu 245 arch/powerpc/kvm/e500_mmu.c int kvmppc_e500_emul_tlbivax(struct kvm_vcpu *vcpu, gva_t ea) kvm_vcpu 307 arch/powerpc/kvm/e500_mmu.c int kvmppc_e500_emul_tlbilx(struct kvm_vcpu *vcpu, int type, gva_t ea) kvm_vcpu 322 arch/powerpc/kvm/e500_mmu.c int kvmppc_e500_emul_tlbre(struct kvm_vcpu *vcpu) kvm_vcpu 341 arch/powerpc/kvm/e500_mmu.c int kvmppc_e500_emul_tlbsx(struct kvm_vcpu *vcpu, gva_t ea) kvm_vcpu 390 arch/powerpc/kvm/e500_mmu.c int kvmppc_e500_emul_tlbwe(struct kvm_vcpu *vcpu) kvm_vcpu 453 arch/powerpc/kvm/e500_mmu.c static int kvmppc_e500_tlb_search(struct kvm_vcpu *vcpu, kvm_vcpu 469 arch/powerpc/kvm/e500_mmu.c int kvmppc_core_vcpu_translate(struct kvm_vcpu *vcpu, kvm_vcpu 495 arch/powerpc/kvm/e500_mmu.c int kvmppc_mmu_itlb_index(struct kvm_vcpu *vcpu, gva_t eaddr) kvm_vcpu 502 arch/powerpc/kvm/e500_mmu.c int kvmppc_mmu_dtlb_index(struct kvm_vcpu *vcpu, gva_t eaddr) kvm_vcpu 509 arch/powerpc/kvm/e500_mmu.c void kvmppc_mmu_itlb_miss(struct kvm_vcpu *vcpu) kvm_vcpu 516 arch/powerpc/kvm/e500_mmu.c void kvmppc_mmu_dtlb_miss(struct kvm_vcpu *vcpu) kvm_vcpu 523 arch/powerpc/kvm/e500_mmu.c gpa_t kvmppc_mmu_xlate(struct kvm_vcpu *vcpu, unsigned int index, kvm_vcpu 536 arch/powerpc/kvm/e500_mmu.c void kvmppc_mmu_destroy_e500(struct kvm_vcpu *vcpu) kvm_vcpu 571 arch/powerpc/kvm/e500_mmu.c void kvmppc_get_sregs_e500_tlb(struct kvm_vcpu *vcpu, struct kvm_sregs *sregs) kvm_vcpu 587 arch/powerpc/kvm/e500_mmu.c int kvmppc_set_sregs_e500_tlb(struct kvm_vcpu *vcpu, struct kvm_sregs *sregs) kvm_vcpu 601 arch/powerpc/kvm/e500_mmu.c int kvmppc_get_one_reg_e500_tlb(struct kvm_vcpu *vcpu, u64 id, kvm_vcpu 654 arch/powerpc/kvm/e500_mmu.c int kvmppc_set_one_reg_e500_tlb(struct kvm_vcpu *vcpu, u64 id, kvm_vcpu 721 arch/powerpc/kvm/e500_mmu.c static int vcpu_mmu_geometry_update(struct kvm_vcpu *vcpu, kvm_vcpu 735 arch/powerpc/kvm/e500_mmu.c int kvm_vcpu_ioctl_config_tlb(struct kvm_vcpu *vcpu, kvm_vcpu 860 arch/powerpc/kvm/e500_mmu.c int kvm_vcpu_ioctl_dirty_tlb(struct kvm_vcpu *vcpu, kvm_vcpu 870 arch/powerpc/kvm/e500_mmu.c static int vcpu_mmu_init(struct kvm_vcpu *vcpu, kvm_vcpu 904 arch/powerpc/kvm/e500_mmu.c struct kvm_vcpu *vcpu = &vcpu_e500->vcpu; kvm_vcpu 158 arch/powerpc/kvm/e500_mmu_host.c void kvmppc_map_magic(struct kvm_vcpu *vcpu) kvm_vcpu 295 arch/powerpc/kvm/e500_mmu_host.c void kvmppc_core_flush_tlb(struct kvm_vcpu *vcpu) kvm_vcpu 305 arch/powerpc/kvm/e500_mmu_host.c struct kvm_vcpu *vcpu, kvm_vcpu 586 arch/powerpc/kvm/e500_mmu_host.c void kvmppc_mmu_map(struct kvm_vcpu *vcpu, u64 eaddr, gpa_t gpaddr, kvm_vcpu 625 arch/powerpc/kvm/e500_mmu_host.c int kvmppc_load_last_inst(struct kvm_vcpu *vcpu, kvm_vcpu 715 arch/powerpc/kvm/e500_mmu_host.c int kvmppc_load_last_inst(struct kvm_vcpu *vcpu, kvm_vcpu 27 arch/powerpc/kvm/e500mc.c void kvmppc_set_pending_interrupt(struct kvm_vcpu *vcpu, enum int_class type) kvm_vcpu 100 arch/powerpc/kvm/e500mc.c void kvmppc_set_pid(struct kvm_vcpu *vcpu, u32 pid) kvm_vcpu 105 arch/powerpc/kvm/e500mc.c void kvmppc_mmu_msr_notify(struct kvm_vcpu *vcpu, u32 old_msr) kvm_vcpu 110 arch/powerpc/kvm/e500mc.c static DEFINE_PER_CPU(struct kvm_vcpu *[KVMPPC_NR_LPIDS], last_vcpu_of_lpid); kvm_vcpu 112 arch/powerpc/kvm/e500mc.c static void kvmppc_core_vcpu_load_e500mc(struct kvm_vcpu *vcpu, int cpu) kvm_vcpu 149 arch/powerpc/kvm/e500mc.c static void kvmppc_core_vcpu_put_e500mc(struct kvm_vcpu *vcpu) kvm_vcpu 195 arch/powerpc/kvm/e500mc.c int kvmppc_core_vcpu_setup(struct kvm_vcpu *vcpu) kvm_vcpu 214 arch/powerpc/kvm/e500mc.c static int kvmppc_core_get_sregs_e500mc(struct kvm_vcpu *vcpu, kvm_vcpu 238 arch/powerpc/kvm/e500mc.c static int kvmppc_core_set_sregs_e500mc(struct kvm_vcpu *vcpu, kvm_vcpu 272 arch/powerpc/kvm/e500mc.c static int kvmppc_get_one_reg_e500mc(struct kvm_vcpu *vcpu, u64 id, kvm_vcpu 288 arch/powerpc/kvm/e500mc.c static int kvmppc_set_one_reg_e500mc(struct kvm_vcpu *vcpu, u64 id, kvm_vcpu 304 arch/powerpc/kvm/e500mc.c static struct kvm_vcpu *kvmppc_core_vcpu_create_e500mc(struct kvm *kvm, kvm_vcpu 308 arch/powerpc/kvm/e500mc.c struct kvm_vcpu *vcpu; kvm_vcpu 348 arch/powerpc/kvm/e500mc.c static void kvmppc_core_vcpu_free_e500mc(struct kvm_vcpu *vcpu) kvm_vcpu 26 arch/powerpc/kvm/emulate.c void kvmppc_emulate_dec(struct kvm_vcpu *vcpu) kvm_vcpu 63 arch/powerpc/kvm/emulate.c u32 kvmppc_get_dec(struct kvm_vcpu *vcpu, u64 tb) kvm_vcpu 75 arch/powerpc/kvm/emulate.c static int kvmppc_emulate_mtspr(struct kvm_vcpu *vcpu, int sprn, int rs) kvm_vcpu 128 arch/powerpc/kvm/emulate.c static int kvmppc_emulate_mfspr(struct kvm_vcpu *vcpu, int sprn, int rt) kvm_vcpu 194 arch/powerpc/kvm/emulate.c int kvmppc_emulate_instruction(struct kvm_run *run, struct kvm_vcpu *vcpu) kvm_vcpu 28 arch/powerpc/kvm/emulate_loadstore.c static bool kvmppc_check_fp_disabled(struct kvm_vcpu *vcpu) kvm_vcpu 40 arch/powerpc/kvm/emulate_loadstore.c static bool kvmppc_check_vsx_disabled(struct kvm_vcpu *vcpu) kvm_vcpu 52 arch/powerpc/kvm/emulate_loadstore.c static bool kvmppc_check_altivec_disabled(struct kvm_vcpu *vcpu) kvm_vcpu 72 arch/powerpc/kvm/emulate_loadstore.c int kvmppc_emulate_loadstore(struct kvm_vcpu *vcpu) kvm_vcpu 117 arch/powerpc/kvm/mpic.c struct kvm_vcpu *vcpu = current->thread.kvm_vcpu; kvm_vcpu 179 arch/powerpc/kvm/mpic.c struct kvm_vcpu *vcpu; kvm_vcpu 1177 arch/powerpc/kvm/mpic.c void kvmppc_mpic_set_epr(struct kvm_vcpu *vcpu) kvm_vcpu 1372 arch/powerpc/kvm/mpic.c static int kvm_mpic_read(struct kvm_vcpu *vcpu, kvm_vcpu 1414 arch/powerpc/kvm/mpic.c static int kvm_mpic_write(struct kvm_vcpu *vcpu, kvm_vcpu 1733 arch/powerpc/kvm/mpic.c int kvmppc_mpic_connect_vcpu(struct kvm_device *dev, struct kvm_vcpu *vcpu, kvm_vcpu 1778 arch/powerpc/kvm/mpic.c void kvmppc_mpic_disconnect_vcpu(struct openpic *opp, struct kvm_vcpu *vcpu) kvm_vcpu 48 arch/powerpc/kvm/powerpc.c int kvm_arch_vcpu_runnable(struct kvm_vcpu *v) kvm_vcpu 53 arch/powerpc/kvm/powerpc.c bool kvm_arch_dy_runnable(struct kvm_vcpu *vcpu) kvm_vcpu 58 arch/powerpc/kvm/powerpc.c bool kvm_arch_vcpu_in_kernel(struct kvm_vcpu *vcpu) kvm_vcpu 63 arch/powerpc/kvm/powerpc.c int kvm_arch_vcpu_should_kick(struct kvm_vcpu *vcpu) kvm_vcpu 77 arch/powerpc/kvm/powerpc.c int kvmppc_prepare_to_enter(struct kvm_vcpu *vcpu) kvm_vcpu 140 arch/powerpc/kvm/powerpc.c static void kvmppc_swab_shared(struct kvm_vcpu *vcpu) kvm_vcpu 160 arch/powerpc/kvm/powerpc.c int kvmppc_kvm_pv(struct kvm_vcpu *vcpu) kvm_vcpu 252 arch/powerpc/kvm/powerpc.c int kvmppc_sanity_check(struct kvm_vcpu *vcpu) kvm_vcpu 281 arch/powerpc/kvm/powerpc.c int kvmppc_emulate_mmio(struct kvm_run *run, struct kvm_vcpu *vcpu) kvm_vcpu 323 arch/powerpc/kvm/powerpc.c int kvmppc_st(struct kvm_vcpu *vcpu, ulong *eaddr, int size, void *ptr, kvm_vcpu 366 arch/powerpc/kvm/powerpc.c int kvmppc_ld(struct kvm_vcpu *vcpu, ulong *eaddr, int size, void *ptr, kvm_vcpu 463 arch/powerpc/kvm/powerpc.c struct kvm_vcpu *vcpu; kvm_vcpu 719 arch/powerpc/kvm/powerpc.c struct kvm_vcpu *kvm_arch_vcpu_create(struct kvm *kvm, unsigned int id) kvm_vcpu 721 arch/powerpc/kvm/powerpc.c struct kvm_vcpu *vcpu; kvm_vcpu 730 arch/powerpc/kvm/powerpc.c void kvm_arch_vcpu_postcreate(struct kvm_vcpu *vcpu) kvm_vcpu 734 arch/powerpc/kvm/powerpc.c void kvm_arch_vcpu_free(struct kvm_vcpu *vcpu) kvm_vcpu 759 arch/powerpc/kvm/powerpc.c void kvm_arch_vcpu_destroy(struct kvm_vcpu *vcpu) kvm_vcpu 764 arch/powerpc/kvm/powerpc.c int kvm_cpu_has_pending_timer(struct kvm_vcpu *vcpu) kvm_vcpu 771 arch/powerpc/kvm/powerpc.c struct kvm_vcpu *vcpu; kvm_vcpu 773 arch/powerpc/kvm/powerpc.c vcpu = container_of(timer, struct kvm_vcpu, arch.dec_timer); kvm_vcpu 779 arch/powerpc/kvm/powerpc.c int kvm_arch_vcpu_init(struct kvm_vcpu *vcpu) kvm_vcpu 794 arch/powerpc/kvm/powerpc.c void kvm_arch_vcpu_uninit(struct kvm_vcpu *vcpu) kvm_vcpu 800 arch/powerpc/kvm/powerpc.c void kvm_arch_vcpu_load(struct kvm_vcpu *vcpu, int cpu) kvm_vcpu 815 arch/powerpc/kvm/powerpc.c void kvm_arch_vcpu_put(struct kvm_vcpu *vcpu) kvm_vcpu 891 arch/powerpc/kvm/powerpc.c static inline void kvmppc_set_vsr_dword(struct kvm_vcpu *vcpu, kvm_vcpu 910 arch/powerpc/kvm/powerpc.c static inline void kvmppc_set_vsr_dword_dump(struct kvm_vcpu *vcpu, kvm_vcpu 927 arch/powerpc/kvm/powerpc.c static inline void kvmppc_set_vsr_word_dump(struct kvm_vcpu *vcpu, kvm_vcpu 947 arch/powerpc/kvm/powerpc.c static inline void kvmppc_set_vsr_word(struct kvm_vcpu *vcpu, kvm_vcpu 973 arch/powerpc/kvm/powerpc.c static inline int kvmppc_get_vmx_offset_generic(struct kvm_vcpu *vcpu, kvm_vcpu 990 arch/powerpc/kvm/powerpc.c static inline int kvmppc_get_vmx_dword_offset(struct kvm_vcpu *vcpu, kvm_vcpu 996 arch/powerpc/kvm/powerpc.c static inline int kvmppc_get_vmx_word_offset(struct kvm_vcpu *vcpu, kvm_vcpu 1002 arch/powerpc/kvm/powerpc.c static inline int kvmppc_get_vmx_hword_offset(struct kvm_vcpu *vcpu, kvm_vcpu 1008 arch/powerpc/kvm/powerpc.c static inline int kvmppc_get_vmx_byte_offset(struct kvm_vcpu *vcpu, kvm_vcpu 1015 arch/powerpc/kvm/powerpc.c static inline void kvmppc_set_vmx_dword(struct kvm_vcpu *vcpu, kvm_vcpu 1031 arch/powerpc/kvm/powerpc.c static inline void kvmppc_set_vmx_word(struct kvm_vcpu *vcpu, kvm_vcpu 1047 arch/powerpc/kvm/powerpc.c static inline void kvmppc_set_vmx_hword(struct kvm_vcpu *vcpu, kvm_vcpu 1063 arch/powerpc/kvm/powerpc.c static inline void kvmppc_set_vmx_byte(struct kvm_vcpu *vcpu, kvm_vcpu 1110 arch/powerpc/kvm/powerpc.c static void kvmppc_complete_mmio_load(struct kvm_vcpu *vcpu, kvm_vcpu 1222 arch/powerpc/kvm/powerpc.c static int __kvmppc_handle_load(struct kvm_run *run, struct kvm_vcpu *vcpu, kvm_vcpu 1267 arch/powerpc/kvm/powerpc.c int kvmppc_handle_load(struct kvm_run *run, struct kvm_vcpu *vcpu, kvm_vcpu 1276 arch/powerpc/kvm/powerpc.c int kvmppc_handle_loads(struct kvm_run *run, struct kvm_vcpu *vcpu, kvm_vcpu 1284 arch/powerpc/kvm/powerpc.c int kvmppc_handle_vsx_load(struct kvm_run *run, struct kvm_vcpu *vcpu, kvm_vcpu 1310 arch/powerpc/kvm/powerpc.c int kvmppc_handle_store(struct kvm_run *run, struct kvm_vcpu *vcpu, kvm_vcpu 1372 arch/powerpc/kvm/powerpc.c static inline int kvmppc_get_vsr_data(struct kvm_vcpu *vcpu, int rs, u64 *val) kvm_vcpu 1426 arch/powerpc/kvm/powerpc.c int kvmppc_handle_vsx_store(struct kvm_run *run, struct kvm_vcpu *vcpu, kvm_vcpu 1457 arch/powerpc/kvm/powerpc.c static int kvmppc_emulate_mmio_vsx_loadstore(struct kvm_vcpu *vcpu, kvm_vcpu 1493 arch/powerpc/kvm/powerpc.c int kvmppc_handle_vmx_load(struct kvm_run *run, struct kvm_vcpu *vcpu, kvm_vcpu 1516 arch/powerpc/kvm/powerpc.c int kvmppc_get_vmx_dword(struct kvm_vcpu *vcpu, int index, u64 *val) kvm_vcpu 1534 arch/powerpc/kvm/powerpc.c int kvmppc_get_vmx_word(struct kvm_vcpu *vcpu, int index, u64 *val) kvm_vcpu 1552 arch/powerpc/kvm/powerpc.c int kvmppc_get_vmx_hword(struct kvm_vcpu *vcpu, int index, u64 *val) kvm_vcpu 1570 arch/powerpc/kvm/powerpc.c int kvmppc_get_vmx_byte(struct kvm_vcpu *vcpu, int index, u64 *val) kvm_vcpu 1588 arch/powerpc/kvm/powerpc.c int kvmppc_handle_vmx_store(struct kvm_run *run, struct kvm_vcpu *vcpu, kvm_vcpu 1636 arch/powerpc/kvm/powerpc.c static int kvmppc_emulate_mmio_vmx_loadstore(struct kvm_vcpu *vcpu, kvm_vcpu 1671 arch/powerpc/kvm/powerpc.c int kvm_vcpu_ioctl_get_one_reg(struct kvm_vcpu *vcpu, struct kvm_one_reg *reg) kvm_vcpu 1719 arch/powerpc/kvm/powerpc.c int kvm_vcpu_ioctl_set_one_reg(struct kvm_vcpu *vcpu, struct kvm_one_reg *reg) kvm_vcpu 1768 arch/powerpc/kvm/powerpc.c int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu, struct kvm_run *run) kvm_vcpu 1843 arch/powerpc/kvm/powerpc.c int kvm_vcpu_ioctl_interrupt(struct kvm_vcpu *vcpu, struct kvm_interrupt *irq) kvm_vcpu 1857 arch/powerpc/kvm/powerpc.c static int kvm_vcpu_ioctl_enable_cap(struct kvm_vcpu *vcpu, kvm_vcpu 1999 arch/powerpc/kvm/powerpc.c int kvm_arch_vcpu_ioctl_get_mpstate(struct kvm_vcpu *vcpu, kvm_vcpu 2005 arch/powerpc/kvm/powerpc.c int kvm_arch_vcpu_ioctl_set_mpstate(struct kvm_vcpu *vcpu, kvm_vcpu 2014 arch/powerpc/kvm/powerpc.c struct kvm_vcpu *vcpu = filp->private_data; kvm_vcpu 2029 arch/powerpc/kvm/powerpc.c struct kvm_vcpu *vcpu = filp->private_data; kvm_vcpu 2080 arch/powerpc/kvm/powerpc.c vm_fault_t kvm_arch_vcpu_fault(struct kvm_vcpu *vcpu, struct vm_fault *vmf) kvm_vcpu 22 arch/powerpc/kvm/timing.c void kvmppc_init_timing_stats(struct kvm_vcpu *vcpu) kvm_vcpu 44 arch/powerpc/kvm/timing.c static void add_exit_timing(struct kvm_vcpu *vcpu, u64 duration, int type) kvm_vcpu 82 arch/powerpc/kvm/timing.c void kvmppc_update_timing_stats(struct kvm_vcpu *vcpu) kvm_vcpu 131 arch/powerpc/kvm/timing.c struct kvm_vcpu *vcpu = m->private; kvm_vcpu 179 arch/powerpc/kvm/timing.c struct kvm_vcpu *vcpu = seqf->private; kvm_vcpu 207 arch/powerpc/kvm/timing.c void kvmppc_create_vcpu_debugfs(struct kvm_vcpu *vcpu, unsigned int id) kvm_vcpu 227 arch/powerpc/kvm/timing.c void kvmppc_remove_vcpu_debugfs(struct kvm_vcpu *vcpu) kvm_vcpu 16 arch/powerpc/kvm/timing.h void kvmppc_init_timing_stats(struct kvm_vcpu *vcpu); kvm_vcpu 17 arch/powerpc/kvm/timing.h void kvmppc_update_timing_stats(struct kvm_vcpu *vcpu); kvm_vcpu 18 arch/powerpc/kvm/timing.h void kvmppc_create_vcpu_debugfs(struct kvm_vcpu *vcpu, unsigned int id); kvm_vcpu 19 arch/powerpc/kvm/timing.h void kvmppc_remove_vcpu_debugfs(struct kvm_vcpu *vcpu); kvm_vcpu 21 arch/powerpc/kvm/timing.h static inline void kvmppc_set_exit_type(struct kvm_vcpu *vcpu, int type) kvm_vcpu 28 arch/powerpc/kvm/timing.h static inline void kvmppc_init_timing_stats(struct kvm_vcpu *vcpu) {} kvm_vcpu 29 arch/powerpc/kvm/timing.h static inline void kvmppc_update_timing_stats(struct kvm_vcpu *vcpu) {} kvm_vcpu 30 arch/powerpc/kvm/timing.h static inline void kvmppc_create_vcpu_debugfs(struct kvm_vcpu *vcpu, kvm_vcpu 32 arch/powerpc/kvm/timing.h static inline void kvmppc_remove_vcpu_debugfs(struct kvm_vcpu *vcpu) {} kvm_vcpu 33 arch/powerpc/kvm/timing.h static inline void kvmppc_set_exit_type(struct kvm_vcpu *vcpu, int type) {} kvm_vcpu 37 arch/powerpc/kvm/timing.h static inline void kvmppc_account_exit_stat(struct kvm_vcpu *vcpu, int type) kvm_vcpu 92 arch/powerpc/kvm/timing.h static inline void kvmppc_account_exit(struct kvm_vcpu *vcpu, int type) kvm_vcpu 101 arch/powerpc/kvm/trace.h TP_PROTO(struct kvm_vcpu *vcpu), kvm_vcpu 39 arch/powerpc/kvm/trace_booke.h TP_PROTO(unsigned int exit_nr, struct kvm_vcpu *vcpu), kvm_vcpu 195 arch/powerpc/kvm/trace_booke.h TP_PROTO(struct kvm_vcpu *vcpu, unsigned int priority), kvm_vcpu 222 arch/powerpc/kvm/trace_hv.h TP_PROTO(struct kvm_vcpu *vcpu), kvm_vcpu 246 arch/powerpc/kvm/trace_hv.h TP_PROTO(struct kvm_vcpu *vcpu), kvm_vcpu 273 arch/powerpc/kvm/trace_hv.h TP_PROTO(struct kvm_vcpu *vcpu, unsigned long *hptep, kvm_vcpu 309 arch/powerpc/kvm/trace_hv.h TP_PROTO(struct kvm_vcpu *vcpu, unsigned long *hptep, long ret), kvm_vcpu 333 arch/powerpc/kvm/trace_hv.h TP_PROTO(struct kvm_vcpu *vcpu), kvm_vcpu 362 arch/powerpc/kvm/trace_hv.h TP_PROTO(struct kvm_vcpu *vcpu, int ret), kvm_vcpu 457 arch/powerpc/kvm/trace_hv.h TP_PROTO(struct kvm_vcpu *vcpu), kvm_vcpu 475 arch/powerpc/kvm/trace_hv.h TP_PROTO(struct kvm_vcpu *vcpu, struct kvm_run *run), kvm_vcpu 13 arch/powerpc/kvm/trace_pr.h TP_PROTO(int r, struct kvm_vcpu *vcpu), kvm_vcpu 122 arch/powerpc/kvm/trace_pr.h TP_PROTO(const char *type, struct kvm_vcpu *vcpu, unsigned long long p1, kvm_vcpu 218 arch/powerpc/kvm/trace_pr.h TP_PROTO(unsigned int exit_nr, struct kvm_vcpu *vcpu), kvm_vcpu 1249 arch/powerpc/mm/book3s64/radix_tlb.c if (paca_ptrs[sib]->kvm_hstate.kvm_vcpu) kvm_vcpu 725 arch/s390/include/asm/kvm_host.h int (*hook)(struct kvm_vcpu *vcpu); kvm_vcpu 893 arch/s390/include/asm/kvm_host.h bool kvm_arch_can_inject_async_page_present(struct kvm_vcpu *vcpu); kvm_vcpu 895 arch/s390/include/asm/kvm_host.h void kvm_arch_async_page_ready(struct kvm_vcpu *vcpu, kvm_vcpu 898 arch/s390/include/asm/kvm_host.h void kvm_arch_async_page_not_present(struct kvm_vcpu *vcpu, kvm_vcpu 901 arch/s390/include/asm/kvm_host.h void kvm_arch_async_page_present(struct kvm_vcpu *vcpu, kvm_vcpu 916 arch/s390/include/asm/kvm_host.h static inline void kvm_arch_vcpu_uninit(struct kvm_vcpu *vcpu) {} kvm_vcpu 917 arch/s390/include/asm/kvm_host.h static inline void kvm_arch_sched_in(struct kvm_vcpu *vcpu, int cpu) {} kvm_vcpu 924 arch/s390/include/asm/kvm_host.h static inline void kvm_arch_vcpu_blocking(struct kvm_vcpu *vcpu) {} kvm_vcpu 925 arch/s390/include/asm/kvm_host.h static inline void kvm_arch_vcpu_unblocking(struct kvm_vcpu *vcpu) {} kvm_vcpu 927 arch/s390/include/asm/kvm_host.h void kvm_arch_vcpu_block_finish(struct kvm_vcpu *vcpu); kvm_vcpu 21 arch/s390/kvm/diag.c static int diag_release_pages(struct kvm_vcpu *vcpu) kvm_vcpu 59 arch/s390/kvm/diag.c static int __diag_page_ref_service(struct kvm_vcpu *vcpu) kvm_vcpu 146 arch/s390/kvm/diag.c static int __diag_time_slice_end(struct kvm_vcpu *vcpu) kvm_vcpu 154 arch/s390/kvm/diag.c static int __diag_time_slice_end_directed(struct kvm_vcpu *vcpu) kvm_vcpu 156 arch/s390/kvm/diag.c struct kvm_vcpu *tcpu; kvm_vcpu 172 arch/s390/kvm/diag.c static int __diag_ipl_functions(struct kvm_vcpu *vcpu) kvm_vcpu 202 arch/s390/kvm/diag.c static int __diag_virtio_hypercall(struct kvm_vcpu *vcpu) kvm_vcpu 238 arch/s390/kvm/diag.c int kvm_s390_handle_diag(struct kvm_vcpu *vcpu) kvm_vcpu 264 arch/s390/kvm/gaccess.c int ipte_lock_held(struct kvm_vcpu *vcpu) kvm_vcpu 277 arch/s390/kvm/gaccess.c static void ipte_lock_simple(struct kvm_vcpu *vcpu) kvm_vcpu 303 arch/s390/kvm/gaccess.c static void ipte_unlock_simple(struct kvm_vcpu *vcpu) kvm_vcpu 324 arch/s390/kvm/gaccess.c static void ipte_lock_siif(struct kvm_vcpu *vcpu) kvm_vcpu 345 arch/s390/kvm/gaccess.c static void ipte_unlock_siif(struct kvm_vcpu *vcpu) kvm_vcpu 363 arch/s390/kvm/gaccess.c void ipte_lock(struct kvm_vcpu *vcpu) kvm_vcpu 371 arch/s390/kvm/gaccess.c void ipte_unlock(struct kvm_vcpu *vcpu) kvm_vcpu 379 arch/s390/kvm/gaccess.c static int ar_translation(struct kvm_vcpu *vcpu, union asce *asce, u8 ar, kvm_vcpu 493 arch/s390/kvm/gaccess.c static int trans_exc(struct kvm_vcpu *vcpu, int code, unsigned long gva, kvm_vcpu 554 arch/s390/kvm/gaccess.c static int get_vcpu_asce(struct kvm_vcpu *vcpu, union asce *asce, kvm_vcpu 614 arch/s390/kvm/gaccess.c static unsigned long guest_translate(struct kvm_vcpu *vcpu, unsigned long gva, kvm_vcpu 784 arch/s390/kvm/gaccess.c static int low_address_protection_enabled(struct kvm_vcpu *vcpu, kvm_vcpu 797 arch/s390/kvm/gaccess.c static int guest_page_range(struct kvm_vcpu *vcpu, unsigned long ga, u8 ar, kvm_vcpu 830 arch/s390/kvm/gaccess.c int access_guest(struct kvm_vcpu *vcpu, unsigned long ga, u8 ar, void *data, kvm_vcpu 875 arch/s390/kvm/gaccess.c int access_guest_real(struct kvm_vcpu *vcpu, unsigned long gra, kvm_vcpu 904 arch/s390/kvm/gaccess.c int guest_translate_address(struct kvm_vcpu *vcpu, unsigned long gva, u8 ar, kvm_vcpu 938 arch/s390/kvm/gaccess.c int check_gva_range(struct kvm_vcpu *vcpu, unsigned long gva, u8 ar, kvm_vcpu 966 arch/s390/kvm/gaccess.c int kvm_s390_check_low_addr_prot_real(struct kvm_vcpu *vcpu, unsigned long gra) kvm_vcpu 1164 arch/s390/kvm/gaccess.c int kvm_s390_shadow_fault(struct kvm_vcpu *vcpu, struct gmap *sg, kvm_vcpu 27 arch/s390/kvm/gaccess.h static inline unsigned long kvm_s390_real_to_abs(struct kvm_vcpu *vcpu, kvm_vcpu 52 arch/s390/kvm/gaccess.h static inline unsigned long kvm_s390_logical_to_effective(struct kvm_vcpu *vcpu, kvm_vcpu 94 arch/s390/kvm/gaccess.h struct kvm_vcpu *__vcpu = (vcpu); \ kvm_vcpu 121 arch/s390/kvm/gaccess.h int write_guest_lc(struct kvm_vcpu *vcpu, unsigned long gra, void *data, kvm_vcpu 147 arch/s390/kvm/gaccess.h int read_guest_lc(struct kvm_vcpu *vcpu, unsigned long gra, void *data, kvm_vcpu 161 arch/s390/kvm/gaccess.h int guest_translate_address(struct kvm_vcpu *vcpu, unsigned long gva, kvm_vcpu 163 arch/s390/kvm/gaccess.h int check_gva_range(struct kvm_vcpu *vcpu, unsigned long gva, u8 ar, kvm_vcpu 166 arch/s390/kvm/gaccess.h int access_guest(struct kvm_vcpu *vcpu, unsigned long ga, u8 ar, void *data, kvm_vcpu 169 arch/s390/kvm/gaccess.h int access_guest_real(struct kvm_vcpu *vcpu, unsigned long gra, kvm_vcpu 218 arch/s390/kvm/gaccess.h int write_guest(struct kvm_vcpu *vcpu, unsigned long ga, u8 ar, void *data, kvm_vcpu 238 arch/s390/kvm/gaccess.h int read_guest(struct kvm_vcpu *vcpu, unsigned long ga, u8 ar, void *data, kvm_vcpu 259 arch/s390/kvm/gaccess.h int read_guest_instr(struct kvm_vcpu *vcpu, unsigned long ga, void *data, kvm_vcpu 282 arch/s390/kvm/gaccess.h int write_guest_abs(struct kvm_vcpu *vcpu, unsigned long gpa, void *data, kvm_vcpu 305 arch/s390/kvm/gaccess.h int read_guest_abs(struct kvm_vcpu *vcpu, unsigned long gpa, void *data, kvm_vcpu 328 arch/s390/kvm/gaccess.h int write_guest_real(struct kvm_vcpu *vcpu, unsigned long gra, void *data, kvm_vcpu 351 arch/s390/kvm/gaccess.h int read_guest_real(struct kvm_vcpu *vcpu, unsigned long gra, void *data, kvm_vcpu 357 arch/s390/kvm/gaccess.h void ipte_lock(struct kvm_vcpu *vcpu); kvm_vcpu 358 arch/s390/kvm/gaccess.h void ipte_unlock(struct kvm_vcpu *vcpu); kvm_vcpu 359 arch/s390/kvm/gaccess.h int ipte_lock_held(struct kvm_vcpu *vcpu); kvm_vcpu 360 arch/s390/kvm/gaccess.h int kvm_s390_check_low_addr_prot_real(struct kvm_vcpu *vcpu, unsigned long gra); kvm_vcpu 362 arch/s390/kvm/gaccess.h int kvm_s390_shadow_fault(struct kvm_vcpu *vcpu, struct gmap *shadow, kvm_vcpu 59 arch/s390/kvm/guestdbg.c static void enable_all_hw_bp(struct kvm_vcpu *vcpu) kvm_vcpu 99 arch/s390/kvm/guestdbg.c static void enable_all_hw_wp(struct kvm_vcpu *vcpu) kvm_vcpu 130 arch/s390/kvm/guestdbg.c void kvm_s390_backup_guest_per_regs(struct kvm_vcpu *vcpu) kvm_vcpu 138 arch/s390/kvm/guestdbg.c void kvm_s390_restore_guest_per_regs(struct kvm_vcpu *vcpu) kvm_vcpu 146 arch/s390/kvm/guestdbg.c void kvm_s390_patch_guest_per_regs(struct kvm_vcpu *vcpu) kvm_vcpu 174 arch/s390/kvm/guestdbg.c static int __import_wp_info(struct kvm_vcpu *vcpu, kvm_vcpu 203 arch/s390/kvm/guestdbg.c int kvm_s390_import_bp_data(struct kvm_vcpu *vcpu, kvm_vcpu 282 arch/s390/kvm/guestdbg.c void kvm_s390_clear_bp_data(struct kvm_vcpu *vcpu) kvm_vcpu 313 arch/s390/kvm/guestdbg.c static struct kvm_hw_bp_info_arch *find_hw_bp(struct kvm_vcpu *vcpu, kvm_vcpu 338 arch/s390/kvm/guestdbg.c static struct kvm_hw_wp_info_arch *any_wp_changed(struct kvm_vcpu *vcpu) kvm_vcpu 371 arch/s390/kvm/guestdbg.c void kvm_s390_prepare_debug_exit(struct kvm_vcpu *vcpu) kvm_vcpu 388 arch/s390/kvm/guestdbg.c static int debug_exit_required(struct kvm_vcpu *vcpu, u8 perc, kvm_vcpu 436 arch/s390/kvm/guestdbg.c static int per_fetched_addr(struct kvm_vcpu *vcpu, unsigned long *addr) kvm_vcpu 497 arch/s390/kvm/guestdbg.c int kvm_s390_handle_per_ifetch_icpt(struct kvm_vcpu *vcpu) kvm_vcpu 537 arch/s390/kvm/guestdbg.c static int filter_guest_per_event(struct kvm_vcpu *vcpu) kvm_vcpu 586 arch/s390/kvm/guestdbg.c int kvm_s390_handle_per_event(struct kvm_vcpu *vcpu) kvm_vcpu 25 arch/s390/kvm/intercept.c u8 kvm_s390_get_ilen(struct kvm_vcpu *vcpu) kvm_vcpu 53 arch/s390/kvm/intercept.c static int handle_stop(struct kvm_vcpu *vcpu) kvm_vcpu 87 arch/s390/kvm/intercept.c static int handle_validity(struct kvm_vcpu *vcpu) kvm_vcpu 102 arch/s390/kvm/intercept.c static int handle_instruction(struct kvm_vcpu *vcpu) kvm_vcpu 139 arch/s390/kvm/intercept.c static int inject_prog_on_prog_intercept(struct kvm_vcpu *vcpu) kvm_vcpu 207 arch/s390/kvm/intercept.c static int handle_itdb(struct kvm_vcpu *vcpu) kvm_vcpu 227 arch/s390/kvm/intercept.c static int handle_prog(struct kvm_vcpu *vcpu) kvm_vcpu 267 arch/s390/kvm/intercept.c static int handle_external_interrupt(struct kvm_vcpu *vcpu) kvm_vcpu 314 arch/s390/kvm/intercept.c static int handle_mvpg_pei(struct kvm_vcpu *vcpu) kvm_vcpu 344 arch/s390/kvm/intercept.c static int handle_partial_execution(struct kvm_vcpu *vcpu) kvm_vcpu 361 arch/s390/kvm/intercept.c int handle_sthyi(struct kvm_vcpu *vcpu) kvm_vcpu 411 arch/s390/kvm/intercept.c static int handle_operexc(struct kvm_vcpu *vcpu) kvm_vcpu 447 arch/s390/kvm/intercept.c int kvm_handle_sie_intercept(struct kvm_vcpu *vcpu) kvm_vcpu 42 arch/s390/kvm/interrupt.c static int sca_ext_call_pending(struct kvm_vcpu *vcpu, int *src_id) kvm_vcpu 74 arch/s390/kvm/interrupt.c static int sca_inject_ext_call(struct kvm_vcpu *vcpu, int src_id) kvm_vcpu 115 arch/s390/kvm/interrupt.c static void sca_clear_ext_call(struct kvm_vcpu *vcpu) kvm_vcpu 144 arch/s390/kvm/interrupt.c int psw_extint_disabled(struct kvm_vcpu *vcpu) kvm_vcpu 149 arch/s390/kvm/interrupt.c static int psw_ioint_disabled(struct kvm_vcpu *vcpu) kvm_vcpu 154 arch/s390/kvm/interrupt.c static int psw_mchk_disabled(struct kvm_vcpu *vcpu) kvm_vcpu 159 arch/s390/kvm/interrupt.c static int psw_interrupts_disabled(struct kvm_vcpu *vcpu) kvm_vcpu 166 arch/s390/kvm/interrupt.c static int ckc_interrupts_enabled(struct kvm_vcpu *vcpu) kvm_vcpu 177 arch/s390/kvm/interrupt.c static int ckc_irq_pending(struct kvm_vcpu *vcpu) kvm_vcpu 191 arch/s390/kvm/interrupt.c static int cpu_timer_interrupts_enabled(struct kvm_vcpu *vcpu) kvm_vcpu 197 arch/s390/kvm/interrupt.c static int cpu_timer_irq_pending(struct kvm_vcpu *vcpu) kvm_vcpu 325 arch/s390/kvm/interrupt.c static inline unsigned long pending_irqs_no_gisa(struct kvm_vcpu *vcpu) kvm_vcpu 331 arch/s390/kvm/interrupt.c static inline unsigned long pending_irqs(struct kvm_vcpu *vcpu) kvm_vcpu 352 arch/s390/kvm/interrupt.c static unsigned long disable_iscs(struct kvm_vcpu *vcpu, kvm_vcpu 364 arch/s390/kvm/interrupt.c static unsigned long deliverable_irqs(struct kvm_vcpu *vcpu) kvm_vcpu 408 arch/s390/kvm/interrupt.c static void __set_cpu_idle(struct kvm_vcpu *vcpu) kvm_vcpu 414 arch/s390/kvm/interrupt.c static void __unset_cpu_idle(struct kvm_vcpu *vcpu) kvm_vcpu 420 arch/s390/kvm/interrupt.c static void __reset_intercept_indicators(struct kvm_vcpu *vcpu) kvm_vcpu 434 arch/s390/kvm/interrupt.c static void set_intercept_indicators_io(struct kvm_vcpu *vcpu) kvm_vcpu 444 arch/s390/kvm/interrupt.c static void set_intercept_indicators_ext(struct kvm_vcpu *vcpu) kvm_vcpu 454 arch/s390/kvm/interrupt.c static void set_intercept_indicators_mchk(struct kvm_vcpu *vcpu) kvm_vcpu 464 arch/s390/kvm/interrupt.c static void set_intercept_indicators_stop(struct kvm_vcpu *vcpu) kvm_vcpu 471 arch/s390/kvm/interrupt.c static void set_intercept_indicators(struct kvm_vcpu *vcpu) kvm_vcpu 479 arch/s390/kvm/interrupt.c static int __must_check __deliver_cpu_timer(struct kvm_vcpu *vcpu) kvm_vcpu 499 arch/s390/kvm/interrupt.c static int __must_check __deliver_ckc(struct kvm_vcpu *vcpu) kvm_vcpu 519 arch/s390/kvm/interrupt.c static int __must_check __deliver_pfault_init(struct kvm_vcpu *vcpu) kvm_vcpu 547 arch/s390/kvm/interrupt.c static int __write_machine_check(struct kvm_vcpu *vcpu, kvm_vcpu 645 arch/s390/kvm/interrupt.c static int __must_check __deliver_machine_check(struct kvm_vcpu *vcpu) kvm_vcpu 696 arch/s390/kvm/interrupt.c static int __must_check __deliver_restart(struct kvm_vcpu *vcpu) kvm_vcpu 714 arch/s390/kvm/interrupt.c static int __must_check __deliver_set_prefix(struct kvm_vcpu *vcpu) kvm_vcpu 734 arch/s390/kvm/interrupt.c static int __must_check __deliver_emergency_signal(struct kvm_vcpu *vcpu) kvm_vcpu 762 arch/s390/kvm/interrupt.c static int __must_check __deliver_external_call(struct kvm_vcpu *vcpu) kvm_vcpu 790 arch/s390/kvm/interrupt.c static int __must_check __deliver_prog(struct kvm_vcpu *vcpu) kvm_vcpu 905 arch/s390/kvm/interrupt.c static int __must_check __deliver_service(struct kvm_vcpu *vcpu) kvm_vcpu 939 arch/s390/kvm/interrupt.c static int __must_check __deliver_pfault_done(struct kvm_vcpu *vcpu) kvm_vcpu 981 arch/s390/kvm/interrupt.c static int __must_check __deliver_virtio(struct kvm_vcpu *vcpu) kvm_vcpu 1027 arch/s390/kvm/interrupt.c static int __do_deliver_io(struct kvm_vcpu *vcpu, struct kvm_s390_io_info *io) kvm_vcpu 1044 arch/s390/kvm/interrupt.c static int __must_check __deliver_io(struct kvm_vcpu *vcpu, kvm_vcpu 1114 arch/s390/kvm/interrupt.c int kvm_s390_ext_call_pending(struct kvm_vcpu *vcpu) kvm_vcpu 1124 arch/s390/kvm/interrupt.c int kvm_s390_vcpu_has_irq(struct kvm_vcpu *vcpu, int exclude_stop) kvm_vcpu 1143 arch/s390/kvm/interrupt.c int kvm_cpu_has_pending_timer(struct kvm_vcpu *vcpu) kvm_vcpu 1148 arch/s390/kvm/interrupt.c static u64 __calculate_sltime(struct kvm_vcpu *vcpu) kvm_vcpu 1180 arch/s390/kvm/interrupt.c int kvm_s390_handle_wait(struct kvm_vcpu *vcpu) kvm_vcpu 1225 arch/s390/kvm/interrupt.c void kvm_s390_vcpu_wakeup(struct kvm_vcpu *vcpu) kvm_vcpu 1239 arch/s390/kvm/interrupt.c struct kvm_vcpu *vcpu; kvm_vcpu 1242 arch/s390/kvm/interrupt.c vcpu = container_of(timer, struct kvm_vcpu, arch.ckc_timer); kvm_vcpu 1255 arch/s390/kvm/interrupt.c void kvm_s390_clear_local_irqs(struct kvm_vcpu *vcpu) kvm_vcpu 1268 arch/s390/kvm/interrupt.c int __must_check kvm_s390_deliver_pending_interrupts(struct kvm_vcpu *vcpu) kvm_vcpu 1349 arch/s390/kvm/interrupt.c static int __inject_prog(struct kvm_vcpu *vcpu, struct kvm_s390_irq *irq) kvm_vcpu 1391 arch/s390/kvm/interrupt.c static int __inject_pfault_init(struct kvm_vcpu *vcpu, struct kvm_s390_irq *irq) kvm_vcpu 1408 arch/s390/kvm/interrupt.c static int __inject_extcall(struct kvm_vcpu *vcpu, struct kvm_s390_irq *irq) kvm_vcpu 1434 arch/s390/kvm/interrupt.c static int __inject_set_prefix(struct kvm_vcpu *vcpu, struct kvm_s390_irq *irq) kvm_vcpu 1454 arch/s390/kvm/interrupt.c static int __inject_sigp_stop(struct kvm_vcpu *vcpu, struct kvm_s390_irq *irq) kvm_vcpu 1480 arch/s390/kvm/interrupt.c static int __inject_sigp_restart(struct kvm_vcpu *vcpu, kvm_vcpu 1493 arch/s390/kvm/interrupt.c static int __inject_sigp_emergency(struct kvm_vcpu *vcpu, kvm_vcpu 1514 arch/s390/kvm/interrupt.c static int __inject_mchk(struct kvm_vcpu *vcpu, struct kvm_s390_irq *irq) kvm_vcpu 1546 arch/s390/kvm/interrupt.c static int __inject_ckc(struct kvm_vcpu *vcpu) kvm_vcpu 1560 arch/s390/kvm/interrupt.c static int __inject_cpu_timer(struct kvm_vcpu *vcpu) kvm_vcpu 1811 arch/s390/kvm/interrupt.c struct kvm_vcpu *dst_vcpu; kvm_vcpu 1978 arch/s390/kvm/interrupt.c int kvm_s390_is_stop_irq_pending(struct kvm_vcpu *vcpu) kvm_vcpu 1985 arch/s390/kvm/interrupt.c void kvm_s390_clear_stop_irq(struct kvm_vcpu *vcpu) kvm_vcpu 1995 arch/s390/kvm/interrupt.c static int do_inject_vcpu(struct kvm_vcpu *vcpu, struct kvm_s390_irq *irq) kvm_vcpu 2040 arch/s390/kvm/interrupt.c int kvm_s390_inject_vcpu(struct kvm_vcpu *vcpu, struct kvm_s390_irq *irq) kvm_vcpu 2600 arch/s390/kvm/interrupt.c struct kvm_vcpu *vcpu; kvm_vcpu 2783 arch/s390/kvm/interrupt.c void kvm_s390_reinject_machine_check(struct kvm_vcpu *vcpu, kvm_vcpu 2847 arch/s390/kvm/interrupt.c int kvm_s390_set_irq_state(struct kvm_vcpu *vcpu, void __user *irqstate, int len) kvm_vcpu 2929 arch/s390/kvm/interrupt.c int kvm_s390_get_irq_state(struct kvm_vcpu *vcpu, __u8 __user *buf, int len) kvm_vcpu 2989 arch/s390/kvm/interrupt.c struct kvm_vcpu *vcpu; kvm_vcpu 59 arch/s390/kvm/kvm-s390.c #define VCPU_STAT(x) offsetof(struct kvm_vcpu, stat.x), KVM_STAT_VCPU kvm_vcpu 270 arch/s390/kvm/kvm-s390.c struct kvm_vcpu *vcpu; kvm_vcpu 607 arch/s390/kvm/kvm-s390.c static void sca_del_vcpu(struct kvm_vcpu *vcpu); kvm_vcpu 655 arch/s390/kvm/kvm-s390.c struct kvm_vcpu *vcpu; kvm_vcpu 899 arch/s390/kvm/kvm-s390.c static void kvm_s390_vcpu_crypto_setup(struct kvm_vcpu *vcpu); kvm_vcpu 903 arch/s390/kvm/kvm-s390.c struct kvm_vcpu *vcpu; kvm_vcpu 990 arch/s390/kvm/kvm-s390.c struct kvm_vcpu *vcpu; kvm_vcpu 2524 arch/s390/kvm/kvm-s390.c void kvm_arch_vcpu_destroy(struct kvm_vcpu *vcpu) kvm_vcpu 2547 arch/s390/kvm/kvm-s390.c struct kvm_vcpu *vcpu; kvm_vcpu 2576 arch/s390/kvm/kvm-s390.c static int __kvm_ucontrol_vcpu_init(struct kvm_vcpu *vcpu) kvm_vcpu 2586 arch/s390/kvm/kvm-s390.c static void sca_del_vcpu(struct kvm_vcpu *vcpu) kvm_vcpu 2605 arch/s390/kvm/kvm-s390.c static void sca_add_vcpu(struct kvm_vcpu *vcpu) kvm_vcpu 2657 arch/s390/kvm/kvm-s390.c struct kvm_vcpu *vcpu; kvm_vcpu 2712 arch/s390/kvm/kvm-s390.c int kvm_arch_vcpu_init(struct kvm_vcpu *vcpu) kvm_vcpu 2746 arch/s390/kvm/kvm-s390.c static void __start_cpu_timer_accounting(struct kvm_vcpu *vcpu) kvm_vcpu 2755 arch/s390/kvm/kvm-s390.c static void __stop_cpu_timer_accounting(struct kvm_vcpu *vcpu) kvm_vcpu 2765 arch/s390/kvm/kvm-s390.c static void __enable_cpu_timer_accounting(struct kvm_vcpu *vcpu) kvm_vcpu 2773 arch/s390/kvm/kvm-s390.c static void __disable_cpu_timer_accounting(struct kvm_vcpu *vcpu) kvm_vcpu 2780 arch/s390/kvm/kvm-s390.c static void enable_cpu_timer_accounting(struct kvm_vcpu *vcpu) kvm_vcpu 2787 arch/s390/kvm/kvm-s390.c static void disable_cpu_timer_accounting(struct kvm_vcpu *vcpu) kvm_vcpu 2795 arch/s390/kvm/kvm-s390.c void kvm_s390_set_cpu_timer(struct kvm_vcpu *vcpu, __u64 cputm) kvm_vcpu 2807 arch/s390/kvm/kvm-s390.c __u64 kvm_s390_get_cpu_timer(struct kvm_vcpu *vcpu) kvm_vcpu 2832 arch/s390/kvm/kvm-s390.c void kvm_arch_vcpu_load(struct kvm_vcpu *vcpu, int cpu) kvm_vcpu 2842 arch/s390/kvm/kvm-s390.c void kvm_arch_vcpu_put(struct kvm_vcpu *vcpu) kvm_vcpu 2853 arch/s390/kvm/kvm-s390.c static void kvm_s390_vcpu_initial_reset(struct kvm_vcpu *vcpu) kvm_vcpu 2880 arch/s390/kvm/kvm-s390.c void kvm_arch_vcpu_postcreate(struct kvm_vcpu *vcpu) kvm_vcpu 2917 arch/s390/kvm/kvm-s390.c static void kvm_s390_vcpu_crypto_setup(struct kvm_vcpu *vcpu) kvm_vcpu 2946 arch/s390/kvm/kvm-s390.c void kvm_s390_vcpu_unsetup_cmma(struct kvm_vcpu *vcpu) kvm_vcpu 2952 arch/s390/kvm/kvm-s390.c int kvm_s390_vcpu_setup_cmma(struct kvm_vcpu *vcpu) kvm_vcpu 2960 arch/s390/kvm/kvm-s390.c static void kvm_s390_vcpu_setup_model(struct kvm_vcpu *vcpu) kvm_vcpu 2969 arch/s390/kvm/kvm-s390.c int kvm_arch_vcpu_setup(struct kvm_vcpu *vcpu) kvm_vcpu 3042 arch/s390/kvm/kvm-s390.c struct kvm_vcpu *kvm_arch_vcpu_create(struct kvm *kvm, kvm_vcpu 3045 arch/s390/kvm/kvm-s390.c struct kvm_vcpu *vcpu; kvm_vcpu 3093 arch/s390/kvm/kvm-s390.c int kvm_arch_vcpu_runnable(struct kvm_vcpu *vcpu) kvm_vcpu 3098 arch/s390/kvm/kvm-s390.c bool kvm_arch_vcpu_in_kernel(struct kvm_vcpu *vcpu) kvm_vcpu 3103 arch/s390/kvm/kvm-s390.c void kvm_s390_vcpu_block(struct kvm_vcpu *vcpu) kvm_vcpu 3109 arch/s390/kvm/kvm-s390.c void kvm_s390_vcpu_unblock(struct kvm_vcpu *vcpu) kvm_vcpu 3114 arch/s390/kvm/kvm-s390.c static void kvm_s390_vcpu_request(struct kvm_vcpu *vcpu) kvm_vcpu 3120 arch/s390/kvm/kvm-s390.c bool kvm_s390_vcpu_sie_inhibited(struct kvm_vcpu *vcpu) kvm_vcpu 3126 arch/s390/kvm/kvm-s390.c static void kvm_s390_vcpu_request_handled(struct kvm_vcpu *vcpu) kvm_vcpu 3135 arch/s390/kvm/kvm-s390.c void exit_sie(struct kvm_vcpu *vcpu) kvm_vcpu 3144 arch/s390/kvm/kvm-s390.c void kvm_s390_sync_request(int req, struct kvm_vcpu *vcpu) kvm_vcpu 3154 arch/s390/kvm/kvm-s390.c struct kvm_vcpu *vcpu; kvm_vcpu 3174 arch/s390/kvm/kvm-s390.c bool kvm_arch_no_poll(struct kvm_vcpu *vcpu) kvm_vcpu 3185 arch/s390/kvm/kvm-s390.c int kvm_arch_vcpu_should_kick(struct kvm_vcpu *vcpu) kvm_vcpu 3192 arch/s390/kvm/kvm-s390.c static int kvm_arch_vcpu_ioctl_get_one_reg(struct kvm_vcpu *vcpu, kvm_vcpu 3241 arch/s390/kvm/kvm-s390.c static int kvm_arch_vcpu_ioctl_set_one_reg(struct kvm_vcpu *vcpu, kvm_vcpu 3294 arch/s390/kvm/kvm-s390.c static int kvm_arch_vcpu_ioctl_initial_reset(struct kvm_vcpu *vcpu) kvm_vcpu 3300 arch/s390/kvm/kvm-s390.c int kvm_arch_vcpu_ioctl_set_regs(struct kvm_vcpu *vcpu, struct kvm_regs *regs) kvm_vcpu 3308 arch/s390/kvm/kvm-s390.c int kvm_arch_vcpu_ioctl_get_regs(struct kvm_vcpu *vcpu, struct kvm_regs *regs) kvm_vcpu 3316 arch/s390/kvm/kvm-s390.c int kvm_arch_vcpu_ioctl_set_sregs(struct kvm_vcpu *vcpu, kvm_vcpu 3328 arch/s390/kvm/kvm-s390.c int kvm_arch_vcpu_ioctl_get_sregs(struct kvm_vcpu *vcpu, kvm_vcpu 3340 arch/s390/kvm/kvm-s390.c int kvm_arch_vcpu_ioctl_set_fpu(struct kvm_vcpu *vcpu, struct kvm_fpu *fpu) kvm_vcpu 3362 arch/s390/kvm/kvm-s390.c int kvm_arch_vcpu_ioctl_get_fpu(struct kvm_vcpu *vcpu, struct kvm_fpu *fpu) kvm_vcpu 3379 arch/s390/kvm/kvm-s390.c static int kvm_arch_vcpu_ioctl_set_initial_psw(struct kvm_vcpu *vcpu, psw_t psw) kvm_vcpu 3392 arch/s390/kvm/kvm-s390.c int kvm_arch_vcpu_ioctl_translate(struct kvm_vcpu *vcpu, kvm_vcpu 3402 arch/s390/kvm/kvm-s390.c int kvm_arch_vcpu_ioctl_set_guest_debug(struct kvm_vcpu *vcpu, kvm_vcpu 3444 arch/s390/kvm/kvm-s390.c int kvm_arch_vcpu_ioctl_get_mpstate(struct kvm_vcpu *vcpu, kvm_vcpu 3459 arch/s390/kvm/kvm-s390.c int kvm_arch_vcpu_ioctl_set_mpstate(struct kvm_vcpu *vcpu, kvm_vcpu 3487 arch/s390/kvm/kvm-s390.c static bool ibs_enabled(struct kvm_vcpu *vcpu) kvm_vcpu 3492 arch/s390/kvm/kvm-s390.c static int kvm_s390_handle_requests(struct kvm_vcpu *vcpu) kvm_vcpu 3575 arch/s390/kvm/kvm-s390.c struct kvm_vcpu *vcpu; kvm_vcpu 3613 arch/s390/kvm/kvm-s390.c long kvm_arch_fault_in_page(struct kvm_vcpu *vcpu, gpa_t gpa, int writable) kvm_vcpu 3619 arch/s390/kvm/kvm-s390.c static void __kvm_inject_pfault_token(struct kvm_vcpu *vcpu, bool start_token, kvm_vcpu 3636 arch/s390/kvm/kvm-s390.c void kvm_arch_async_page_not_present(struct kvm_vcpu *vcpu, kvm_vcpu 3643 arch/s390/kvm/kvm-s390.c void kvm_arch_async_page_present(struct kvm_vcpu *vcpu, kvm_vcpu 3650 arch/s390/kvm/kvm-s390.c void kvm_arch_async_page_ready(struct kvm_vcpu *vcpu, kvm_vcpu 3656 arch/s390/kvm/kvm-s390.c bool kvm_arch_can_inject_async_page_present(struct kvm_vcpu *vcpu) kvm_vcpu 3665 arch/s390/kvm/kvm-s390.c static int kvm_arch_setup_async_pf(struct kvm_vcpu *vcpu) kvm_vcpu 3694 arch/s390/kvm/kvm-s390.c static int vcpu_pre_run(struct kvm_vcpu *vcpu) kvm_vcpu 3739 arch/s390/kvm/kvm-s390.c static int vcpu_post_run_fault_in_sie(struct kvm_vcpu *vcpu) kvm_vcpu 3775 arch/s390/kvm/kvm-s390.c static int vcpu_post_run(struct kvm_vcpu *vcpu, int exit_reason) kvm_vcpu 3828 arch/s390/kvm/kvm-s390.c static int __vcpu_run(struct kvm_vcpu *vcpu) kvm_vcpu 3867 arch/s390/kvm/kvm-s390.c static void sync_regs(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run) kvm_vcpu 3959 arch/s390/kvm/kvm-s390.c static void store_regs(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run) kvm_vcpu 3997 arch/s390/kvm/kvm-s390.c int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run) kvm_vcpu 4065 arch/s390/kvm/kvm-s390.c int kvm_s390_store_status_unloaded(struct kvm_vcpu *vcpu, unsigned long gpa) kvm_vcpu 4117 arch/s390/kvm/kvm-s390.c int kvm_s390_vcpu_store_status(struct kvm_vcpu *vcpu, unsigned long addr) kvm_vcpu 4131 arch/s390/kvm/kvm-s390.c static void __disable_ibs_on_vcpu(struct kvm_vcpu *vcpu) kvm_vcpu 4140 arch/s390/kvm/kvm-s390.c struct kvm_vcpu *vcpu; kvm_vcpu 4147 arch/s390/kvm/kvm-s390.c static void __enable_ibs_on_vcpu(struct kvm_vcpu *vcpu) kvm_vcpu 4155 arch/s390/kvm/kvm-s390.c void kvm_s390_vcpu_start(struct kvm_vcpu *vcpu) kvm_vcpu 4194 arch/s390/kvm/kvm-s390.c void kvm_s390_vcpu_stop(struct kvm_vcpu *vcpu) kvm_vcpu 4197 arch/s390/kvm/kvm-s390.c struct kvm_vcpu *started_vcpu = NULL; kvm_vcpu 4232 arch/s390/kvm/kvm-s390.c static int kvm_vcpu_ioctl_enable_cap(struct kvm_vcpu *vcpu, kvm_vcpu 4256 arch/s390/kvm/kvm-s390.c static long kvm_s390_guest_mem_op(struct kvm_vcpu *vcpu, kvm_vcpu 4320 arch/s390/kvm/kvm-s390.c struct kvm_vcpu *vcpu = filp->private_data; kvm_vcpu 4348 arch/s390/kvm/kvm-s390.c struct kvm_vcpu *vcpu = filp->private_data; kvm_vcpu 4485 arch/s390/kvm/kvm-s390.c vm_fault_t kvm_arch_vcpu_fault(struct kvm_vcpu *vcpu, struct vm_fault *vmf) kvm_vcpu 4567 arch/s390/kvm/kvm-s390.c void kvm_arch_vcpu_block_finish(struct kvm_vcpu *vcpu) kvm_vcpu 4590 arch/s390/kvm/kvm-s390.c return kvm_init(NULL, sizeof(struct kvm_vcpu), 0, THIS_MODULE); kvm_vcpu 48 arch/s390/kvm/kvm-s390.h static inline void kvm_s390_set_cpuflags(struct kvm_vcpu *vcpu, u32 flags) kvm_vcpu 53 arch/s390/kvm/kvm-s390.h static inline void kvm_s390_clear_cpuflags(struct kvm_vcpu *vcpu, u32 flags) kvm_vcpu 58 arch/s390/kvm/kvm-s390.h static inline bool kvm_s390_test_cpuflags(struct kvm_vcpu *vcpu, u32 flags) kvm_vcpu 63 arch/s390/kvm/kvm-s390.h static inline int is_vcpu_stopped(struct kvm_vcpu *vcpu) kvm_vcpu 68 arch/s390/kvm/kvm-s390.h static inline int is_vcpu_idle(struct kvm_vcpu *vcpu) kvm_vcpu 85 arch/s390/kvm/kvm-s390.h static inline u32 kvm_s390_get_prefix(struct kvm_vcpu *vcpu) kvm_vcpu 90 arch/s390/kvm/kvm-s390.h static inline void kvm_s390_set_prefix(struct kvm_vcpu *vcpu, u32 prefix) kvm_vcpu 99 arch/s390/kvm/kvm-s390.h static inline u64 kvm_s390_get_base_disp_s(struct kvm_vcpu *vcpu, u8 *ar) kvm_vcpu 110 arch/s390/kvm/kvm-s390.h static inline void kvm_s390_get_base_disp_sse(struct kvm_vcpu *vcpu, kvm_vcpu 128 arch/s390/kvm/kvm-s390.h static inline void kvm_s390_get_regs_rre(struct kvm_vcpu *vcpu, int *r1, int *r2) kvm_vcpu 136 arch/s390/kvm/kvm-s390.h static inline u64 kvm_s390_get_base_disp_rsy(struct kvm_vcpu *vcpu, u8 *ar) kvm_vcpu 151 arch/s390/kvm/kvm-s390.h static inline u64 kvm_s390_get_base_disp_rs(struct kvm_vcpu *vcpu, u8 *ar) kvm_vcpu 163 arch/s390/kvm/kvm-s390.h static inline void kvm_s390_set_psw_cc(struct kvm_vcpu *vcpu, unsigned long cc) kvm_vcpu 200 arch/s390/kvm/kvm-s390.h int kvm_s390_handle_wait(struct kvm_vcpu *vcpu); kvm_vcpu 201 arch/s390/kvm/kvm-s390.h void kvm_s390_vcpu_wakeup(struct kvm_vcpu *vcpu); kvm_vcpu 203 arch/s390/kvm/kvm-s390.h int __must_check kvm_s390_deliver_pending_interrupts(struct kvm_vcpu *vcpu); kvm_vcpu 204 arch/s390/kvm/kvm-s390.h void kvm_s390_clear_local_irqs(struct kvm_vcpu *vcpu); kvm_vcpu 208 arch/s390/kvm/kvm-s390.h int __must_check kvm_s390_inject_vcpu(struct kvm_vcpu *vcpu, kvm_vcpu 210 arch/s390/kvm/kvm-s390.h static inline int kvm_s390_inject_prog_irq(struct kvm_vcpu *vcpu, kvm_vcpu 220 arch/s390/kvm/kvm-s390.h static inline int kvm_s390_inject_program_int(struct kvm_vcpu *vcpu, u16 code) kvm_vcpu 236 arch/s390/kvm/kvm-s390.h u8 kvm_s390_get_ilen(struct kvm_vcpu *vcpu); kvm_vcpu 237 arch/s390/kvm/kvm-s390.h int kvm_handle_sie_intercept(struct kvm_vcpu *vcpu); kvm_vcpu 238 arch/s390/kvm/kvm-s390.h static inline void kvm_s390_rewind_psw(struct kvm_vcpu *vcpu, int ilen) kvm_vcpu 244 arch/s390/kvm/kvm-s390.h static inline void kvm_s390_forward_psw(struct kvm_vcpu *vcpu, int ilen) kvm_vcpu 248 arch/s390/kvm/kvm-s390.h static inline void kvm_s390_retry_instr(struct kvm_vcpu *vcpu) kvm_vcpu 255 arch/s390/kvm/kvm-s390.h int handle_sthyi(struct kvm_vcpu *vcpu); kvm_vcpu 259 arch/s390/kvm/kvm-s390.h int kvm_s390_handle_aa(struct kvm_vcpu *vcpu); kvm_vcpu 260 arch/s390/kvm/kvm-s390.h int kvm_s390_handle_b2(struct kvm_vcpu *vcpu); kvm_vcpu 261 arch/s390/kvm/kvm-s390.h int kvm_s390_handle_e3(struct kvm_vcpu *vcpu); kvm_vcpu 262 arch/s390/kvm/kvm-s390.h int kvm_s390_handle_e5(struct kvm_vcpu *vcpu); kvm_vcpu 263 arch/s390/kvm/kvm-s390.h int kvm_s390_handle_01(struct kvm_vcpu *vcpu); kvm_vcpu 264 arch/s390/kvm/kvm-s390.h int kvm_s390_handle_b9(struct kvm_vcpu *vcpu); kvm_vcpu 265 arch/s390/kvm/kvm-s390.h int kvm_s390_handle_lpsw(struct kvm_vcpu *vcpu); kvm_vcpu 266 arch/s390/kvm/kvm-s390.h int kvm_s390_handle_stctl(struct kvm_vcpu *vcpu); kvm_vcpu 267 arch/s390/kvm/kvm-s390.h int kvm_s390_handle_lctl(struct kvm_vcpu *vcpu); kvm_vcpu 268 arch/s390/kvm/kvm-s390.h int kvm_s390_handle_eb(struct kvm_vcpu *vcpu); kvm_vcpu 269 arch/s390/kvm/kvm-s390.h int kvm_s390_skey_check_enable(struct kvm_vcpu *vcpu); kvm_vcpu 272 arch/s390/kvm/kvm-s390.h int kvm_s390_handle_vsie(struct kvm_vcpu *vcpu); kvm_vcpu 273 arch/s390/kvm/kvm-s390.h void kvm_s390_vsie_kick(struct kvm_vcpu *vcpu); kvm_vcpu 280 arch/s390/kvm/kvm-s390.h int kvm_s390_handle_sigp(struct kvm_vcpu *vcpu); kvm_vcpu 281 arch/s390/kvm/kvm-s390.h int kvm_s390_handle_sigp_pei(struct kvm_vcpu *vcpu); kvm_vcpu 286 arch/s390/kvm/kvm-s390.h long kvm_arch_fault_in_page(struct kvm_vcpu *vcpu, gpa_t gpa, int writable); kvm_vcpu 287 arch/s390/kvm/kvm-s390.h int kvm_s390_store_status_unloaded(struct kvm_vcpu *vcpu, unsigned long addr); kvm_vcpu 288 arch/s390/kvm/kvm-s390.h int kvm_s390_vcpu_store_status(struct kvm_vcpu *vcpu, unsigned long addr); kvm_vcpu 289 arch/s390/kvm/kvm-s390.h void kvm_s390_vcpu_start(struct kvm_vcpu *vcpu); kvm_vcpu 290 arch/s390/kvm/kvm-s390.h void kvm_s390_vcpu_stop(struct kvm_vcpu *vcpu); kvm_vcpu 291 arch/s390/kvm/kvm-s390.h void kvm_s390_vcpu_block(struct kvm_vcpu *vcpu); kvm_vcpu 292 arch/s390/kvm/kvm-s390.h void kvm_s390_vcpu_unblock(struct kvm_vcpu *vcpu); kvm_vcpu 293 arch/s390/kvm/kvm-s390.h bool kvm_s390_vcpu_sie_inhibited(struct kvm_vcpu *vcpu); kvm_vcpu 294 arch/s390/kvm/kvm-s390.h void exit_sie(struct kvm_vcpu *vcpu); kvm_vcpu 295 arch/s390/kvm/kvm-s390.h void kvm_s390_sync_request(int req, struct kvm_vcpu *vcpu); kvm_vcpu 296 arch/s390/kvm/kvm-s390.h int kvm_s390_vcpu_setup_cmma(struct kvm_vcpu *vcpu); kvm_vcpu 297 arch/s390/kvm/kvm-s390.h void kvm_s390_vcpu_unsetup_cmma(struct kvm_vcpu *vcpu); kvm_vcpu 298 arch/s390/kvm/kvm-s390.h void kvm_s390_set_cpu_timer(struct kvm_vcpu *vcpu, __u64 cputm); kvm_vcpu 299 arch/s390/kvm/kvm-s390.h __u64 kvm_s390_get_cpu_timer(struct kvm_vcpu *vcpu); kvm_vcpu 302 arch/s390/kvm/kvm-s390.h int kvm_s390_handle_diag(struct kvm_vcpu *vcpu); kvm_vcpu 307 arch/s390/kvm/kvm-s390.h struct kvm_vcpu *vcpu; kvm_vcpu 317 arch/s390/kvm/kvm-s390.h struct kvm_vcpu *vcpu; kvm_vcpu 359 arch/s390/kvm/kvm-s390.h static inline int kvm_s390_inject_prog_cond(struct kvm_vcpu *vcpu, int rc) kvm_vcpu 370 arch/s390/kvm/kvm-s390.h int kvm_s390_vcpu_has_irq(struct kvm_vcpu *vcpu, int exclude_stop); kvm_vcpu 371 arch/s390/kvm/kvm-s390.h int psw_extint_disabled(struct kvm_vcpu *vcpu); kvm_vcpu 373 arch/s390/kvm/kvm-s390.h int kvm_s390_ext_call_pending(struct kvm_vcpu *vcpu); kvm_vcpu 375 arch/s390/kvm/kvm-s390.h int kvm_s390_is_stop_irq_pending(struct kvm_vcpu *vcpu); kvm_vcpu 376 arch/s390/kvm/kvm-s390.h void kvm_s390_clear_stop_irq(struct kvm_vcpu *vcpu); kvm_vcpu 377 arch/s390/kvm/kvm-s390.h int kvm_s390_set_irq_state(struct kvm_vcpu *vcpu, kvm_vcpu 379 arch/s390/kvm/kvm-s390.h int kvm_s390_get_irq_state(struct kvm_vcpu *vcpu, kvm_vcpu 388 arch/s390/kvm/kvm-s390.h void kvm_s390_backup_guest_per_regs(struct kvm_vcpu *vcpu); kvm_vcpu 389 arch/s390/kvm/kvm-s390.h void kvm_s390_restore_guest_per_regs(struct kvm_vcpu *vcpu); kvm_vcpu 390 arch/s390/kvm/kvm-s390.h void kvm_s390_patch_guest_per_regs(struct kvm_vcpu *vcpu); kvm_vcpu 391 arch/s390/kvm/kvm-s390.h int kvm_s390_import_bp_data(struct kvm_vcpu *vcpu, kvm_vcpu 393 arch/s390/kvm/kvm-s390.h void kvm_s390_clear_bp_data(struct kvm_vcpu *vcpu); kvm_vcpu 394 arch/s390/kvm/kvm-s390.h void kvm_s390_prepare_debug_exit(struct kvm_vcpu *vcpu); kvm_vcpu 395 arch/s390/kvm/kvm-s390.h int kvm_s390_handle_per_ifetch_icpt(struct kvm_vcpu *vcpu); kvm_vcpu 396 arch/s390/kvm/kvm-s390.h int kvm_s390_handle_per_event(struct kvm_vcpu *vcpu); kvm_vcpu 414 arch/s390/kvm/kvm-s390.h void kvm_s390_reinject_machine_check(struct kvm_vcpu *vcpu, kvm_vcpu 35 arch/s390/kvm/priv.c static int handle_ri(struct kvm_vcpu *vcpu) kvm_vcpu 48 arch/s390/kvm/priv.c int kvm_s390_handle_aa(struct kvm_vcpu *vcpu) kvm_vcpu 56 arch/s390/kvm/priv.c static int handle_gs(struct kvm_vcpu *vcpu) kvm_vcpu 76 arch/s390/kvm/priv.c int kvm_s390_handle_e3(struct kvm_vcpu *vcpu) kvm_vcpu 86 arch/s390/kvm/priv.c static int handle_set_clock(struct kvm_vcpu *vcpu) kvm_vcpu 112 arch/s390/kvm/priv.c static int handle_set_prefix(struct kvm_vcpu *vcpu) kvm_vcpu 150 arch/s390/kvm/priv.c static int handle_store_prefix(struct kvm_vcpu *vcpu) kvm_vcpu 180 arch/s390/kvm/priv.c static int handle_store_cpu_address(struct kvm_vcpu *vcpu) kvm_vcpu 206 arch/s390/kvm/priv.c int kvm_s390_skey_check_enable(struct kvm_vcpu *vcpu) kvm_vcpu 230 arch/s390/kvm/priv.c static int try_handle_skey(struct kvm_vcpu *vcpu) kvm_vcpu 246 arch/s390/kvm/priv.c static int handle_iske(struct kvm_vcpu *vcpu) kvm_vcpu 294 arch/s390/kvm/priv.c static int handle_rrbe(struct kvm_vcpu *vcpu) kvm_vcpu 343 arch/s390/kvm/priv.c static int handle_sske(struct kvm_vcpu *vcpu) kvm_vcpu 427 arch/s390/kvm/priv.c static int handle_ipte_interlock(struct kvm_vcpu *vcpu) kvm_vcpu 438 arch/s390/kvm/priv.c static int handle_test_block(struct kvm_vcpu *vcpu) kvm_vcpu 468 arch/s390/kvm/priv.c static int handle_tpi(struct kvm_vcpu *vcpu) kvm_vcpu 534 arch/s390/kvm/priv.c static int handle_tsch(struct kvm_vcpu *vcpu) kvm_vcpu 567 arch/s390/kvm/priv.c static int handle_io_inst(struct kvm_vcpu *vcpu) kvm_vcpu 611 arch/s390/kvm/priv.c static int handle_pqap(struct kvm_vcpu *vcpu) kvm_vcpu 681 arch/s390/kvm/priv.c static int handle_stfl(struct kvm_vcpu *vcpu) kvm_vcpu 727 arch/s390/kvm/priv.c int kvm_s390_handle_lpsw(struct kvm_vcpu *vcpu) kvm_vcpu 757 arch/s390/kvm/priv.c static int handle_lpswe(struct kvm_vcpu *vcpu) kvm_vcpu 781 arch/s390/kvm/priv.c static int handle_stidp(struct kvm_vcpu *vcpu) kvm_vcpu 806 arch/s390/kvm/priv.c static void handle_stsi_3_2_2(struct kvm_vcpu *vcpu, struct sysinfo_3_2_2 *mem) kvm_vcpu 833 arch/s390/kvm/priv.c static void insert_stsi_usr_data(struct kvm_vcpu *vcpu, u64 addr, u8 ar, kvm_vcpu 844 arch/s390/kvm/priv.c static int handle_stsi(struct kvm_vcpu *vcpu) kvm_vcpu 920 arch/s390/kvm/priv.c int kvm_s390_handle_b2(struct kvm_vcpu *vcpu) kvm_vcpu 978 arch/s390/kvm/priv.c static int handle_epsw(struct kvm_vcpu *vcpu) kvm_vcpu 1007 arch/s390/kvm/priv.c static int handle_pfmf(struct kvm_vcpu *vcpu) kvm_vcpu 1122 arch/s390/kvm/priv.c static inline int __do_essa(struct kvm_vcpu *vcpu, const int orc) kvm_vcpu 1184 arch/s390/kvm/priv.c static int handle_essa(struct kvm_vcpu *vcpu) kvm_vcpu 1256 arch/s390/kvm/priv.c int kvm_s390_handle_b9(struct kvm_vcpu *vcpu) kvm_vcpu 1274 arch/s390/kvm/priv.c int kvm_s390_handle_lctl(struct kvm_vcpu *vcpu) kvm_vcpu 1313 arch/s390/kvm/priv.c int kvm_s390_handle_stctl(struct kvm_vcpu *vcpu) kvm_vcpu 1347 arch/s390/kvm/priv.c static int handle_lctlg(struct kvm_vcpu *vcpu) kvm_vcpu 1385 arch/s390/kvm/priv.c static int handle_stctg(struct kvm_vcpu *vcpu) kvm_vcpu 1419 arch/s390/kvm/priv.c int kvm_s390_handle_eb(struct kvm_vcpu *vcpu) kvm_vcpu 1435 arch/s390/kvm/priv.c static int handle_tprot(struct kvm_vcpu *vcpu) kvm_vcpu 1490 arch/s390/kvm/priv.c int kvm_s390_handle_e5(struct kvm_vcpu *vcpu) kvm_vcpu 1500 arch/s390/kvm/priv.c static int handle_sckpf(struct kvm_vcpu *vcpu) kvm_vcpu 1519 arch/s390/kvm/priv.c static int handle_ptff(struct kvm_vcpu *vcpu) kvm_vcpu 1528 arch/s390/kvm/priv.c int kvm_s390_handle_01(struct kvm_vcpu *vcpu) kvm_vcpu 20 arch/s390/kvm/sigp.c static int __sigp_sense(struct kvm_vcpu *vcpu, struct kvm_vcpu *dst_vcpu, kvm_vcpu 44 arch/s390/kvm/sigp.c static int __inject_sigp_emergency(struct kvm_vcpu *vcpu, kvm_vcpu 45 arch/s390/kvm/sigp.c struct kvm_vcpu *dst_vcpu) kvm_vcpu 61 arch/s390/kvm/sigp.c static int __sigp_emergency(struct kvm_vcpu *vcpu, struct kvm_vcpu *dst_vcpu) kvm_vcpu 66 arch/s390/kvm/sigp.c static int __sigp_conditional_emergency(struct kvm_vcpu *vcpu, kvm_vcpu 67 arch/s390/kvm/sigp.c struct kvm_vcpu *dst_vcpu, kvm_vcpu 93 arch/s390/kvm/sigp.c static int __sigp_external_call(struct kvm_vcpu *vcpu, kvm_vcpu 94 arch/s390/kvm/sigp.c struct kvm_vcpu *dst_vcpu, u64 *reg) kvm_vcpu 115 arch/s390/kvm/sigp.c static int __sigp_stop(struct kvm_vcpu *vcpu, struct kvm_vcpu *dst_vcpu) kvm_vcpu 132 arch/s390/kvm/sigp.c static int __sigp_stop_and_store_status(struct kvm_vcpu *vcpu, kvm_vcpu 133 arch/s390/kvm/sigp.c struct kvm_vcpu *dst_vcpu, u64 *reg) kvm_vcpu 151 arch/s390/kvm/sigp.c static int __sigp_set_arch(struct kvm_vcpu *vcpu, u32 parameter, kvm_vcpu 155 arch/s390/kvm/sigp.c struct kvm_vcpu *v; kvm_vcpu 173 arch/s390/kvm/sigp.c static int __sigp_set_prefix(struct kvm_vcpu *vcpu, struct kvm_vcpu *dst_vcpu, kvm_vcpu 203 arch/s390/kvm/sigp.c static int __sigp_store_status_at_addr(struct kvm_vcpu *vcpu, kvm_vcpu 204 arch/s390/kvm/sigp.c struct kvm_vcpu *dst_vcpu, kvm_vcpu 225 arch/s390/kvm/sigp.c static int __sigp_sense_running(struct kvm_vcpu *vcpu, kvm_vcpu 226 arch/s390/kvm/sigp.c struct kvm_vcpu *dst_vcpu, u64 *reg) kvm_vcpu 252 arch/s390/kvm/sigp.c static int __prepare_sigp_re_start(struct kvm_vcpu *vcpu, kvm_vcpu 253 arch/s390/kvm/sigp.c struct kvm_vcpu *dst_vcpu, u8 order_code) kvm_vcpu 268 arch/s390/kvm/sigp.c static int __prepare_sigp_cpu_reset(struct kvm_vcpu *vcpu, kvm_vcpu 269 arch/s390/kvm/sigp.c struct kvm_vcpu *dst_vcpu, u8 order_code) kvm_vcpu 275 arch/s390/kvm/sigp.c static int __prepare_sigp_unknown(struct kvm_vcpu *vcpu, kvm_vcpu 276 arch/s390/kvm/sigp.c struct kvm_vcpu *dst_vcpu) kvm_vcpu 282 arch/s390/kvm/sigp.c static int handle_sigp_dst(struct kvm_vcpu *vcpu, u8 order_code, kvm_vcpu 286 arch/s390/kvm/sigp.c struct kvm_vcpu *dst_vcpu = kvm_get_vcpu_by_id(vcpu->kvm, cpu_addr); kvm_vcpu 359 arch/s390/kvm/sigp.c static int handle_sigp_order_in_user_space(struct kvm_vcpu *vcpu, u8 order_code, kvm_vcpu 409 arch/s390/kvm/sigp.c int kvm_s390_handle_sigp(struct kvm_vcpu *vcpu) kvm_vcpu 460 arch/s390/kvm/sigp.c int kvm_s390_handle_sigp_pei(struct kvm_vcpu *vcpu) kvm_vcpu 464 arch/s390/kvm/sigp.c struct kvm_vcpu *dest_vcpu; kvm_vcpu 43 arch/s390/kvm/trace-s390.h TP_PROTO(unsigned int id, struct kvm_vcpu *vcpu, kvm_vcpu 49 arch/s390/kvm/trace-s390.h __field(struct kvm_vcpu *, vcpu) kvm_vcpu 20 arch/s390/kvm/trace.h #define VCPU_PROTO_COMMON struct kvm_vcpu *vcpu kvm_vcpu 102 arch/s390/kvm/vsie.c static int prepare_cpuflags(struct kvm_vcpu *vcpu, struct vsie_page *vsie_page) kvm_vcpu 139 arch/s390/kvm/vsie.c static int setup_apcb10(struct kvm_vcpu *vcpu, struct kvm_s390_apcb1 *apcb_s, kvm_vcpu 164 arch/s390/kvm/vsie.c static int setup_apcb00(struct kvm_vcpu *vcpu, unsigned long *apcb_s, kvm_vcpu 185 arch/s390/kvm/vsie.c static int setup_apcb11(struct kvm_vcpu *vcpu, unsigned long *apcb_s, kvm_vcpu 212 arch/s390/kvm/vsie.c static int setup_apcb(struct kvm_vcpu *vcpu, struct kvm_s390_crypto_cb *crycb_s, kvm_vcpu 283 arch/s390/kvm/vsie.c static int shadow_crycb(struct kvm_vcpu *vcpu, struct vsie_page *vsie_page) kvm_vcpu 358 arch/s390/kvm/vsie.c static void prepare_ibc(struct kvm_vcpu *vcpu, struct vsie_page *vsie_page) kvm_vcpu 381 arch/s390/kvm/vsie.c static void unshadow_scb(struct kvm_vcpu *vcpu, struct vsie_page *vsie_page) kvm_vcpu 437 arch/s390/kvm/vsie.c static int shadow_scb(struct kvm_vcpu *vcpu, struct vsie_page *vsie_page) kvm_vcpu 606 arch/s390/kvm/vsie.c static int map_prefix(struct kvm_vcpu *vcpu, struct vsie_page *vsie_page) kvm_vcpu 663 arch/s390/kvm/vsie.c static void unpin_blocks(struct kvm_vcpu *vcpu, struct vsie_page *vsie_page) kvm_vcpu 719 arch/s390/kvm/vsie.c static int pin_blocks(struct kvm_vcpu *vcpu, struct vsie_page *vsie_page) kvm_vcpu 837 arch/s390/kvm/vsie.c static void unpin_scb(struct kvm_vcpu *vcpu, struct vsie_page *vsie_page, kvm_vcpu 853 arch/s390/kvm/vsie.c static int pin_scb(struct kvm_vcpu *vcpu, struct vsie_page *vsie_page, kvm_vcpu 875 arch/s390/kvm/vsie.c static int inject_fault(struct kvm_vcpu *vcpu, __u16 code, __u64 vaddr, kvm_vcpu 905 arch/s390/kvm/vsie.c static int handle_fault(struct kvm_vcpu *vcpu, struct vsie_page *vsie_page) kvm_vcpu 932 arch/s390/kvm/vsie.c static void handle_last_fault(struct kvm_vcpu *vcpu, kvm_vcpu 969 arch/s390/kvm/vsie.c static int handle_stfle(struct kvm_vcpu *vcpu, struct vsie_page *vsie_page) kvm_vcpu 992 arch/s390/kvm/vsie.c static int do_vsie_run(struct kvm_vcpu *vcpu, struct vsie_page *vsie_page) kvm_vcpu 1087 arch/s390/kvm/vsie.c static int acquire_gmap_shadow(struct kvm_vcpu *vcpu, kvm_vcpu 1121 arch/s390/kvm/vsie.c static void register_shadow_scb(struct kvm_vcpu *vcpu, kvm_vcpu 1151 arch/s390/kvm/vsie.c static void unregister_shadow_scb(struct kvm_vcpu *vcpu) kvm_vcpu 1165 arch/s390/kvm/vsie.c static int vsie_run(struct kvm_vcpu *vcpu, struct vsie_page *vsie_page) kvm_vcpu 1285 arch/s390/kvm/vsie.c int kvm_s390_handle_vsie(struct kvm_vcpu *vcpu) kvm_vcpu 1366 arch/s390/kvm/vsie.c void kvm_s390_vsie_kick(struct kvm_vcpu *vcpu) kvm_vcpu 380 arch/x86/include/asm/kvm_host.h void (*set_cr3)(struct kvm_vcpu *vcpu, unsigned long root); kvm_vcpu 381 arch/x86/include/asm/kvm_host.h unsigned long (*get_cr3)(struct kvm_vcpu *vcpu); kvm_vcpu 382 arch/x86/include/asm/kvm_host.h u64 (*get_pdptr)(struct kvm_vcpu *vcpu, int index); kvm_vcpu 383 arch/x86/include/asm/kvm_host.h int (*page_fault)(struct kvm_vcpu *vcpu, gpa_t cr2_or_gpa, u32 err, kvm_vcpu 385 arch/x86/include/asm/kvm_host.h void (*inject_page_fault)(struct kvm_vcpu *vcpu, kvm_vcpu 387 arch/x86/include/asm/kvm_host.h gpa_t (*gva_to_gpa)(struct kvm_vcpu *vcpu, gpa_t gva_or_gpa, kvm_vcpu 389 arch/x86/include/asm/kvm_host.h gpa_t (*translate_gpa)(struct kvm_vcpu *vcpu, gpa_t gpa, u32 access, kvm_vcpu 391 arch/x86/include/asm/kvm_host.h int (*sync_page)(struct kvm_vcpu *vcpu, kvm_vcpu 393 arch/x86/include/asm/kvm_host.h void (*invlpg)(struct kvm_vcpu *vcpu, gva_t gva, hpa_t root_hpa); kvm_vcpu 394 arch/x86/include/asm/kvm_host.h void (*update_pte)(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp, kvm_vcpu 456 arch/x86/include/asm/kvm_host.h struct kvm_vcpu *vcpu; kvm_vcpu 660 arch/x86/include/asm/kvm_host.h int (*complete_userspace_io)(struct kvm_vcpu *vcpu); kvm_vcpu 1018 arch/x86/include/asm/kvm_host.h void (*cpuid_update)(struct kvm_vcpu *vcpu); kvm_vcpu 1026 arch/x86/include/asm/kvm_host.h struct kvm_vcpu *(*vcpu_create)(struct kvm *kvm, unsigned id); kvm_vcpu 1027 arch/x86/include/asm/kvm_host.h void (*vcpu_free)(struct kvm_vcpu *vcpu); kvm_vcpu 1028 arch/x86/include/asm/kvm_host.h void (*vcpu_reset)(struct kvm_vcpu *vcpu, bool init_event); kvm_vcpu 1030 arch/x86/include/asm/kvm_host.h void (*prepare_guest_switch)(struct kvm_vcpu *vcpu); kvm_vcpu 1031 arch/x86/include/asm/kvm_host.h void (*vcpu_load)(struct kvm_vcpu *vcpu, int cpu); kvm_vcpu 1032 arch/x86/include/asm/kvm_host.h void (*vcpu_put)(struct kvm_vcpu *vcpu); kvm_vcpu 1034 arch/x86/include/asm/kvm_host.h void (*update_bp_intercept)(struct kvm_vcpu *vcpu); kvm_vcpu 1035 arch/x86/include/asm/kvm_host.h int (*get_msr)(struct kvm_vcpu *vcpu, struct msr_data *msr); kvm_vcpu 1036 arch/x86/include/asm/kvm_host.h int (*set_msr)(struct kvm_vcpu *vcpu, struct msr_data *msr); kvm_vcpu 1037 arch/x86/include/asm/kvm_host.h u64 (*get_segment_base)(struct kvm_vcpu *vcpu, int seg); kvm_vcpu 1038 arch/x86/include/asm/kvm_host.h void (*get_segment)(struct kvm_vcpu *vcpu, kvm_vcpu 1040 arch/x86/include/asm/kvm_host.h int (*get_cpl)(struct kvm_vcpu *vcpu); kvm_vcpu 1041 arch/x86/include/asm/kvm_host.h void (*set_segment)(struct kvm_vcpu *vcpu, kvm_vcpu 1043 arch/x86/include/asm/kvm_host.h void (*get_cs_db_l_bits)(struct kvm_vcpu *vcpu, int *db, int *l); kvm_vcpu 1044 arch/x86/include/asm/kvm_host.h void (*decache_cr0_guest_bits)(struct kvm_vcpu *vcpu); kvm_vcpu 1045 arch/x86/include/asm/kvm_host.h void (*decache_cr3)(struct kvm_vcpu *vcpu); kvm_vcpu 1046 arch/x86/include/asm/kvm_host.h void (*decache_cr4_guest_bits)(struct kvm_vcpu *vcpu); kvm_vcpu 1047 arch/x86/include/asm/kvm_host.h void (*set_cr0)(struct kvm_vcpu *vcpu, unsigned long cr0); kvm_vcpu 1048 arch/x86/include/asm/kvm_host.h void (*set_cr3)(struct kvm_vcpu *vcpu, unsigned long cr3); kvm_vcpu 1049 arch/x86/include/asm/kvm_host.h int (*set_cr4)(struct kvm_vcpu *vcpu, unsigned long cr4); kvm_vcpu 1050 arch/x86/include/asm/kvm_host.h void (*set_efer)(struct kvm_vcpu *vcpu, u64 efer); kvm_vcpu 1051 arch/x86/include/asm/kvm_host.h void (*get_idt)(struct kvm_vcpu *vcpu, struct desc_ptr *dt); kvm_vcpu 1052 arch/x86/include/asm/kvm_host.h void (*set_idt)(struct kvm_vcpu *vcpu, struct desc_ptr *dt); kvm_vcpu 1053 arch/x86/include/asm/kvm_host.h void (*get_gdt)(struct kvm_vcpu *vcpu, struct desc_ptr *dt); kvm_vcpu 1054 arch/x86/include/asm/kvm_host.h void (*set_gdt)(struct kvm_vcpu *vcpu, struct desc_ptr *dt); kvm_vcpu 1055 arch/x86/include/asm/kvm_host.h u64 (*get_dr6)(struct kvm_vcpu *vcpu); kvm_vcpu 1056 arch/x86/include/asm/kvm_host.h void (*set_dr6)(struct kvm_vcpu *vcpu, unsigned long value); kvm_vcpu 1057 arch/x86/include/asm/kvm_host.h void (*sync_dirty_debug_regs)(struct kvm_vcpu *vcpu); kvm_vcpu 1058 arch/x86/include/asm/kvm_host.h void (*set_dr7)(struct kvm_vcpu *vcpu, unsigned long value); kvm_vcpu 1059 arch/x86/include/asm/kvm_host.h void (*cache_reg)(struct kvm_vcpu *vcpu, enum kvm_reg reg); kvm_vcpu 1060 arch/x86/include/asm/kvm_host.h unsigned long (*get_rflags)(struct kvm_vcpu *vcpu); kvm_vcpu 1061 arch/x86/include/asm/kvm_host.h void (*set_rflags)(struct kvm_vcpu *vcpu, unsigned long rflags); kvm_vcpu 1063 arch/x86/include/asm/kvm_host.h void (*tlb_flush)(struct kvm_vcpu *vcpu, bool invalidate_gpa); kvm_vcpu 1074 arch/x86/include/asm/kvm_host.h void (*tlb_flush_gva)(struct kvm_vcpu *vcpu, gva_t addr); kvm_vcpu 1076 arch/x86/include/asm/kvm_host.h void (*run)(struct kvm_vcpu *vcpu); kvm_vcpu 1077 arch/x86/include/asm/kvm_host.h int (*handle_exit)(struct kvm_vcpu *vcpu); kvm_vcpu 1078 arch/x86/include/asm/kvm_host.h int (*skip_emulated_instruction)(struct kvm_vcpu *vcpu); kvm_vcpu 1079 arch/x86/include/asm/kvm_host.h void (*set_interrupt_shadow)(struct kvm_vcpu *vcpu, int mask); kvm_vcpu 1080 arch/x86/include/asm/kvm_host.h u32 (*get_interrupt_shadow)(struct kvm_vcpu *vcpu); kvm_vcpu 1081 arch/x86/include/asm/kvm_host.h void (*patch_hypercall)(struct kvm_vcpu *vcpu, kvm_vcpu 1083 arch/x86/include/asm/kvm_host.h void (*set_irq)(struct kvm_vcpu *vcpu); kvm_vcpu 1084 arch/x86/include/asm/kvm_host.h void (*set_nmi)(struct kvm_vcpu *vcpu); kvm_vcpu 1085 arch/x86/include/asm/kvm_host.h void (*queue_exception)(struct kvm_vcpu *vcpu); kvm_vcpu 1086 arch/x86/include/asm/kvm_host.h void (*cancel_injection)(struct kvm_vcpu *vcpu); kvm_vcpu 1087 arch/x86/include/asm/kvm_host.h int (*interrupt_allowed)(struct kvm_vcpu *vcpu); kvm_vcpu 1088 arch/x86/include/asm/kvm_host.h int (*nmi_allowed)(struct kvm_vcpu *vcpu); kvm_vcpu 1089 arch/x86/include/asm/kvm_host.h bool (*get_nmi_mask)(struct kvm_vcpu *vcpu); kvm_vcpu 1090 arch/x86/include/asm/kvm_host.h void (*set_nmi_mask)(struct kvm_vcpu *vcpu, bool masked); kvm_vcpu 1091 arch/x86/include/asm/kvm_host.h void (*enable_nmi_window)(struct kvm_vcpu *vcpu); kvm_vcpu 1092 arch/x86/include/asm/kvm_host.h void (*enable_irq_window)(struct kvm_vcpu *vcpu); kvm_vcpu 1093 arch/x86/include/asm/kvm_host.h void (*update_cr8_intercept)(struct kvm_vcpu *vcpu, int tpr, int irr); kvm_vcpu 1094 arch/x86/include/asm/kvm_host.h bool (*get_enable_apicv)(struct kvm_vcpu *vcpu); kvm_vcpu 1095 arch/x86/include/asm/kvm_host.h void (*refresh_apicv_exec_ctrl)(struct kvm_vcpu *vcpu); kvm_vcpu 1096 arch/x86/include/asm/kvm_host.h void (*hwapic_irr_update)(struct kvm_vcpu *vcpu, int max_irr); kvm_vcpu 1097 arch/x86/include/asm/kvm_host.h void (*hwapic_isr_update)(struct kvm_vcpu *vcpu, int isr); kvm_vcpu 1098 arch/x86/include/asm/kvm_host.h bool (*guest_apic_has_interrupt)(struct kvm_vcpu *vcpu); kvm_vcpu 1099 arch/x86/include/asm/kvm_host.h void (*load_eoi_exitmap)(struct kvm_vcpu *vcpu, u64 *eoi_exit_bitmap); kvm_vcpu 1100 arch/x86/include/asm/kvm_host.h void (*set_virtual_apic_mode)(struct kvm_vcpu *vcpu); kvm_vcpu 1101 arch/x86/include/asm/kvm_host.h void (*set_apic_access_page_addr)(struct kvm_vcpu *vcpu, hpa_t hpa); kvm_vcpu 1102 arch/x86/include/asm/kvm_host.h int (*deliver_posted_interrupt)(struct kvm_vcpu *vcpu, int vector); kvm_vcpu 1103 arch/x86/include/asm/kvm_host.h int (*sync_pir_to_irr)(struct kvm_vcpu *vcpu); kvm_vcpu 1106 arch/x86/include/asm/kvm_host.h int (*get_tdp_level)(struct kvm_vcpu *vcpu); kvm_vcpu 1107 arch/x86/include/asm/kvm_host.h u64 (*get_mt_mask)(struct kvm_vcpu *vcpu, gfn_t gfn, bool is_mmio); kvm_vcpu 1112 arch/x86/include/asm/kvm_host.h void (*set_tdp_cr3)(struct kvm_vcpu *vcpu, unsigned long cr3); kvm_vcpu 1118 arch/x86/include/asm/kvm_host.h u64 (*read_l1_tsc_offset)(struct kvm_vcpu *vcpu); kvm_vcpu 1120 arch/x86/include/asm/kvm_host.h u64 (*write_l1_tsc_offset)(struct kvm_vcpu *vcpu, u64 offset); kvm_vcpu 1122 arch/x86/include/asm/kvm_host.h void (*get_exit_info)(struct kvm_vcpu *vcpu, u64 *info1, u64 *info2); kvm_vcpu 1124 arch/x86/include/asm/kvm_host.h int (*check_intercept)(struct kvm_vcpu *vcpu, kvm_vcpu 1127 arch/x86/include/asm/kvm_host.h void (*handle_exit_irqoff)(struct kvm_vcpu *vcpu); kvm_vcpu 1134 arch/x86/include/asm/kvm_host.h int (*check_nested_events)(struct kvm_vcpu *vcpu); kvm_vcpu 1135 arch/x86/include/asm/kvm_host.h void (*request_immediate_exit)(struct kvm_vcpu *vcpu); kvm_vcpu 1137 arch/x86/include/asm/kvm_host.h void (*sched_in)(struct kvm_vcpu *kvm, int cpu); kvm_vcpu 1163 arch/x86/include/asm/kvm_host.h int (*write_log_dirty)(struct kvm_vcpu *vcpu); kvm_vcpu 1177 arch/x86/include/asm/kvm_host.h int (*pre_block)(struct kvm_vcpu *vcpu); kvm_vcpu 1178 arch/x86/include/asm/kvm_host.h void (*post_block)(struct kvm_vcpu *vcpu); kvm_vcpu 1180 arch/x86/include/asm/kvm_host.h void (*vcpu_blocking)(struct kvm_vcpu *vcpu); kvm_vcpu 1181 arch/x86/include/asm/kvm_host.h void (*vcpu_unblocking)(struct kvm_vcpu *vcpu); kvm_vcpu 1185 arch/x86/include/asm/kvm_host.h void (*apicv_post_state_restore)(struct kvm_vcpu *vcpu); kvm_vcpu 1186 arch/x86/include/asm/kvm_host.h bool (*dy_apicv_has_pending_interrupt)(struct kvm_vcpu *vcpu); kvm_vcpu 1188 arch/x86/include/asm/kvm_host.h int (*set_hv_timer)(struct kvm_vcpu *vcpu, u64 guest_deadline_tsc, kvm_vcpu 1190 arch/x86/include/asm/kvm_host.h void (*cancel_hv_timer)(struct kvm_vcpu *vcpu); kvm_vcpu 1192 arch/x86/include/asm/kvm_host.h void (*setup_mce)(struct kvm_vcpu *vcpu); kvm_vcpu 1194 arch/x86/include/asm/kvm_host.h int (*get_nested_state)(struct kvm_vcpu *vcpu, kvm_vcpu 1197 arch/x86/include/asm/kvm_host.h int (*set_nested_state)(struct kvm_vcpu *vcpu, kvm_vcpu 1200 arch/x86/include/asm/kvm_host.h bool (*get_vmcs12_pages)(struct kvm_vcpu *vcpu); kvm_vcpu 1202 arch/x86/include/asm/kvm_host.h int (*smi_allowed)(struct kvm_vcpu *vcpu); kvm_vcpu 1203 arch/x86/include/asm/kvm_host.h int (*pre_enter_smm)(struct kvm_vcpu *vcpu, char *smstate); kvm_vcpu 1204 arch/x86/include/asm/kvm_host.h int (*pre_leave_smm)(struct kvm_vcpu *vcpu, const char *smstate); kvm_vcpu 1205 arch/x86/include/asm/kvm_host.h int (*enable_smi_window)(struct kvm_vcpu *vcpu); kvm_vcpu 1213 arch/x86/include/asm/kvm_host.h int (*nested_enable_evmcs)(struct kvm_vcpu *vcpu, kvm_vcpu 1215 arch/x86/include/asm/kvm_host.h uint16_t (*nested_get_evmcs_version)(struct kvm_vcpu *vcpu); kvm_vcpu 1217 arch/x86/include/asm/kvm_host.h bool (*need_emulation_on_page_fault)(struct kvm_vcpu *vcpu); kvm_vcpu 1219 arch/x86/include/asm/kvm_host.h bool (*apic_init_signal_blocked)(struct kvm_vcpu *vcpu); kvm_vcpu 1220 arch/x86/include/asm/kvm_host.h int (*enable_direct_tlbflush)(struct kvm_vcpu *vcpu); kvm_vcpu 1257 arch/x86/include/asm/kvm_host.h void kvm_mmu_destroy(struct kvm_vcpu *vcpu); kvm_vcpu 1258 arch/x86/include/asm/kvm_host.h int kvm_mmu_create(struct kvm_vcpu *vcpu); kvm_vcpu 1265 arch/x86/include/asm/kvm_host.h void kvm_mmu_reset_context(struct kvm_vcpu *vcpu); kvm_vcpu 1284 arch/x86/include/asm/kvm_host.h int load_pdptrs(struct kvm_vcpu *vcpu, struct kvm_mmu *mmu, unsigned long cr3); kvm_vcpu 1285 arch/x86/include/asm/kvm_host.h bool pdptrs_changed(struct kvm_vcpu *vcpu); kvm_vcpu 1287 arch/x86/include/asm/kvm_host.h int emulator_write_phys(struct kvm_vcpu *vcpu, gpa_t gpa, kvm_vcpu 1305 arch/x86/include/asm/kvm_host.h u64 vcpu_tsc_khz(struct kvm_vcpu *vcpu); kvm_vcpu 1356 arch/x86/include/asm/kvm_host.h int kvm_emulate_instruction(struct kvm_vcpu *vcpu, int emulation_type); kvm_vcpu 1357 arch/x86/include/asm/kvm_host.h int kvm_emulate_instruction_from_buffer(struct kvm_vcpu *vcpu, kvm_vcpu 1361 arch/x86/include/asm/kvm_host.h bool kvm_valid_efer(struct kvm_vcpu *vcpu, u64 efer); kvm_vcpu 1362 arch/x86/include/asm/kvm_host.h int kvm_get_msr(struct kvm_vcpu *vcpu, u32 index, u64 *data); kvm_vcpu 1363 arch/x86/include/asm/kvm_host.h int kvm_set_msr(struct kvm_vcpu *vcpu, u32 index, u64 data); kvm_vcpu 1364 arch/x86/include/asm/kvm_host.h int kvm_emulate_rdmsr(struct kvm_vcpu *vcpu); kvm_vcpu 1365 arch/x86/include/asm/kvm_host.h int kvm_emulate_wrmsr(struct kvm_vcpu *vcpu); kvm_vcpu 1369 arch/x86/include/asm/kvm_host.h int kvm_fast_pio(struct kvm_vcpu *vcpu, int size, unsigned short port, int in); kvm_vcpu 1370 arch/x86/include/asm/kvm_host.h int kvm_emulate_cpuid(struct kvm_vcpu *vcpu); kvm_vcpu 1371 arch/x86/include/asm/kvm_host.h int kvm_emulate_halt(struct kvm_vcpu *vcpu); kvm_vcpu 1372 arch/x86/include/asm/kvm_host.h int kvm_vcpu_halt(struct kvm_vcpu *vcpu); kvm_vcpu 1373 arch/x86/include/asm/kvm_host.h int kvm_emulate_wbinvd(struct kvm_vcpu *vcpu); kvm_vcpu 1375 arch/x86/include/asm/kvm_host.h void kvm_get_segment(struct kvm_vcpu *vcpu, struct kvm_segment *var, int seg); kvm_vcpu 1376 arch/x86/include/asm/kvm_host.h int kvm_load_segment_descriptor(struct kvm_vcpu *vcpu, u16 selector, int seg); kvm_vcpu 1377 arch/x86/include/asm/kvm_host.h void kvm_vcpu_deliver_sipi_vector(struct kvm_vcpu *vcpu, u8 vector); kvm_vcpu 1379 arch/x86/include/asm/kvm_host.h int kvm_task_switch(struct kvm_vcpu *vcpu, u16 tss_selector, int idt_index, kvm_vcpu 1382 arch/x86/include/asm/kvm_host.h int kvm_set_cr0(struct kvm_vcpu *vcpu, unsigned long cr0); kvm_vcpu 1383 arch/x86/include/asm/kvm_host.h int kvm_set_cr3(struct kvm_vcpu *vcpu, unsigned long cr3); kvm_vcpu 1384 arch/x86/include/asm/kvm_host.h int kvm_set_cr4(struct kvm_vcpu *vcpu, unsigned long cr4); kvm_vcpu 1385 arch/x86/include/asm/kvm_host.h int kvm_set_cr8(struct kvm_vcpu *vcpu, unsigned long cr8); kvm_vcpu 1386 arch/x86/include/asm/kvm_host.h int kvm_set_dr(struct kvm_vcpu *vcpu, int dr, unsigned long val); kvm_vcpu 1387 arch/x86/include/asm/kvm_host.h int kvm_get_dr(struct kvm_vcpu *vcpu, int dr, unsigned long *val); kvm_vcpu 1388 arch/x86/include/asm/kvm_host.h unsigned long kvm_get_cr8(struct kvm_vcpu *vcpu); kvm_vcpu 1389 arch/x86/include/asm/kvm_host.h void kvm_lmsw(struct kvm_vcpu *vcpu, unsigned long msw); kvm_vcpu 1390 arch/x86/include/asm/kvm_host.h void kvm_get_cs_db_l_bits(struct kvm_vcpu *vcpu, int *db, int *l); kvm_vcpu 1391 arch/x86/include/asm/kvm_host.h int kvm_set_xcr(struct kvm_vcpu *vcpu, u32 index, u64 xcr); kvm_vcpu 1393 arch/x86/include/asm/kvm_host.h int kvm_get_msr_common(struct kvm_vcpu *vcpu, struct msr_data *msr); kvm_vcpu 1394 arch/x86/include/asm/kvm_host.h int kvm_set_msr_common(struct kvm_vcpu *vcpu, struct msr_data *msr); kvm_vcpu 1396 arch/x86/include/asm/kvm_host.h unsigned long kvm_get_rflags(struct kvm_vcpu *vcpu); kvm_vcpu 1397 arch/x86/include/asm/kvm_host.h void kvm_set_rflags(struct kvm_vcpu *vcpu, unsigned long rflags); kvm_vcpu 1398 arch/x86/include/asm/kvm_host.h bool kvm_rdpmc(struct kvm_vcpu *vcpu); kvm_vcpu 1400 arch/x86/include/asm/kvm_host.h void kvm_queue_exception(struct kvm_vcpu *vcpu, unsigned nr); kvm_vcpu 1401 arch/x86/include/asm/kvm_host.h void kvm_queue_exception_e(struct kvm_vcpu *vcpu, unsigned nr, u32 error_code); kvm_vcpu 1402 arch/x86/include/asm/kvm_host.h void kvm_requeue_exception(struct kvm_vcpu *vcpu, unsigned nr); kvm_vcpu 1403 arch/x86/include/asm/kvm_host.h void kvm_requeue_exception_e(struct kvm_vcpu *vcpu, unsigned nr, u32 error_code); kvm_vcpu 1404 arch/x86/include/asm/kvm_host.h void kvm_inject_page_fault(struct kvm_vcpu *vcpu, struct x86_exception *fault); kvm_vcpu 1405 arch/x86/include/asm/kvm_host.h int kvm_read_guest_page_mmu(struct kvm_vcpu *vcpu, struct kvm_mmu *mmu, kvm_vcpu 1408 arch/x86/include/asm/kvm_host.h bool kvm_require_cpl(struct kvm_vcpu *vcpu, int required_cpl); kvm_vcpu 1409 arch/x86/include/asm/kvm_host.h bool kvm_require_dr(struct kvm_vcpu *vcpu, int dr); kvm_vcpu 1430 arch/x86/include/asm/kvm_host.h void kvm_inject_nmi(struct kvm_vcpu *vcpu); kvm_vcpu 1433 arch/x86/include/asm/kvm_host.h int kvm_mmu_unprotect_page_virt(struct kvm_vcpu *vcpu, gva_t gva); kvm_vcpu 1434 arch/x86/include/asm/kvm_host.h void __kvm_mmu_free_some_pages(struct kvm_vcpu *vcpu); kvm_vcpu 1435 arch/x86/include/asm/kvm_host.h int kvm_mmu_load(struct kvm_vcpu *vcpu); kvm_vcpu 1436 arch/x86/include/asm/kvm_host.h void kvm_mmu_unload(struct kvm_vcpu *vcpu); kvm_vcpu 1437 arch/x86/include/asm/kvm_host.h void kvm_mmu_sync_roots(struct kvm_vcpu *vcpu); kvm_vcpu 1438 arch/x86/include/asm/kvm_host.h void kvm_mmu_free_roots(struct kvm_vcpu *vcpu, struct kvm_mmu *mmu, kvm_vcpu 1440 arch/x86/include/asm/kvm_host.h gpa_t translate_nested_gpa(struct kvm_vcpu *vcpu, gpa_t gpa, u32 access, kvm_vcpu 1442 arch/x86/include/asm/kvm_host.h gpa_t kvm_mmu_gva_to_gpa_read(struct kvm_vcpu *vcpu, gva_t gva, kvm_vcpu 1444 arch/x86/include/asm/kvm_host.h gpa_t kvm_mmu_gva_to_gpa_fetch(struct kvm_vcpu *vcpu, gva_t gva, kvm_vcpu 1446 arch/x86/include/asm/kvm_host.h gpa_t kvm_mmu_gva_to_gpa_write(struct kvm_vcpu *vcpu, gva_t gva, kvm_vcpu 1448 arch/x86/include/asm/kvm_host.h gpa_t kvm_mmu_gva_to_gpa_system(struct kvm_vcpu *vcpu, gva_t gva, kvm_vcpu 1451 arch/x86/include/asm/kvm_host.h void kvm_vcpu_deactivate_apicv(struct kvm_vcpu *vcpu); kvm_vcpu 1453 arch/x86/include/asm/kvm_host.h int kvm_emulate_hypercall(struct kvm_vcpu *vcpu); kvm_vcpu 1455 arch/x86/include/asm/kvm_host.h int kvm_mmu_page_fault(struct kvm_vcpu *vcpu, gpa_t cr2_or_gpa, u64 error_code, kvm_vcpu 1457 arch/x86/include/asm/kvm_host.h void kvm_mmu_invlpg(struct kvm_vcpu *vcpu, gva_t gva); kvm_vcpu 1458 arch/x86/include/asm/kvm_host.h void kvm_mmu_invpcid_gva(struct kvm_vcpu *vcpu, gva_t gva, unsigned long pcid); kvm_vcpu 1459 arch/x86/include/asm/kvm_host.h void kvm_mmu_new_cr3(struct kvm_vcpu *vcpu, gpa_t new_cr3, bool skip_tlb_flush); kvm_vcpu 1464 arch/x86/include/asm/kvm_host.h static inline gpa_t translate_gpa(struct kvm_vcpu *vcpu, gpa_t gpa, u32 access, kvm_vcpu 1504 arch/x86/include/asm/kvm_host.h static inline void kvm_inject_gp(struct kvm_vcpu *vcpu, u32 error_code) kvm_vcpu 1560 arch/x86/include/asm/kvm_host.h int kvm_cpu_has_injectable_intr(struct kvm_vcpu *v); kvm_vcpu 1561 arch/x86/include/asm/kvm_host.h int kvm_cpu_has_interrupt(struct kvm_vcpu *vcpu); kvm_vcpu 1562 arch/x86/include/asm/kvm_host.h int kvm_arch_interrupt_allowed(struct kvm_vcpu *vcpu); kvm_vcpu 1563 arch/x86/include/asm/kvm_host.h int kvm_cpu_get_interrupt(struct kvm_vcpu *v); kvm_vcpu 1564 arch/x86/include/asm/kvm_host.h void kvm_vcpu_reset(struct kvm_vcpu *vcpu, bool init_event); kvm_vcpu 1565 arch/x86/include/asm/kvm_host.h void kvm_vcpu_reload_apic_access_page(struct kvm_vcpu *vcpu); kvm_vcpu 1574 arch/x86/include/asm/kvm_host.h u64 kvm_scale_tsc(struct kvm_vcpu *vcpu, u64 tsc); kvm_vcpu 1575 arch/x86/include/asm/kvm_host.h u64 kvm_read_l1_tsc(struct kvm_vcpu *vcpu, u64 host_tsc); kvm_vcpu 1577 arch/x86/include/asm/kvm_host.h unsigned long kvm_get_linear_rip(struct kvm_vcpu *vcpu); kvm_vcpu 1578 arch/x86/include/asm/kvm_host.h bool kvm_is_linear_rip(struct kvm_vcpu *vcpu, unsigned long linear_rip); kvm_vcpu 1583 arch/x86/include/asm/kvm_host.h void kvm_arch_async_page_not_present(struct kvm_vcpu *vcpu, kvm_vcpu 1585 arch/x86/include/asm/kvm_host.h void kvm_arch_async_page_present(struct kvm_vcpu *vcpu, kvm_vcpu 1587 arch/x86/include/asm/kvm_host.h void kvm_arch_async_page_ready(struct kvm_vcpu *vcpu, kvm_vcpu 1589 arch/x86/include/asm/kvm_host.h bool kvm_arch_can_inject_async_page_present(struct kvm_vcpu *vcpu); kvm_vcpu 1590 arch/x86/include/asm/kvm_host.h extern bool kvm_find_async_pf_gfn(struct kvm_vcpu *vcpu, gfn_t gfn); kvm_vcpu 1592 arch/x86/include/asm/kvm_host.h int kvm_skip_emulated_instruction(struct kvm_vcpu *vcpu); kvm_vcpu 1593 arch/x86/include/asm/kvm_host.h int kvm_complete_insn_gp(struct kvm_vcpu *vcpu, int err); kvm_vcpu 1594 arch/x86/include/asm/kvm_host.h void __kvm_request_immediate_exit(struct kvm_vcpu *vcpu); kvm_vcpu 1600 arch/x86/include/asm/kvm_host.h bool kvm_vcpu_is_reset_bsp(struct kvm_vcpu *vcpu); kvm_vcpu 1601 arch/x86/include/asm/kvm_host.h bool kvm_vcpu_is_bsp(struct kvm_vcpu *vcpu); kvm_vcpu 1604 arch/x86/include/asm/kvm_host.h struct kvm_vcpu **dest_vcpu); kvm_vcpu 1616 arch/x86/include/asm/kvm_host.h static inline void kvm_arch_vcpu_blocking(struct kvm_vcpu *vcpu) kvm_vcpu 1622 arch/x86/include/asm/kvm_host.h static inline void kvm_arch_vcpu_unblocking(struct kvm_vcpu *vcpu) kvm_vcpu 1628 arch/x86/include/asm/kvm_host.h static inline void kvm_arch_vcpu_block_finish(struct kvm_vcpu *vcpu) {} kvm_vcpu 35 arch/x86/include/asm/kvm_page_track.h void (*track_write)(struct kvm_vcpu *vcpu, gpa_t gpa, const u8 *new, kvm_vcpu 63 arch/x86/include/asm/kvm_page_track.h bool kvm_page_track_is_active(struct kvm_vcpu *vcpu, gfn_t gfn, kvm_vcpu 72 arch/x86/include/asm/kvm_page_track.h void kvm_page_track_write(struct kvm_vcpu *vcpu, gpa_t gpa, const u8 *new, kvm_vcpu 67 arch/x86/kvm/cpuid.c int kvm_update_cpuid(struct kvm_vcpu *vcpu) kvm_vcpu 163 arch/x86/kvm/cpuid.c static void cpuid_fix_nx_cap(struct kvm_vcpu *vcpu) kvm_vcpu 182 arch/x86/kvm/cpuid.c int cpuid_query_maxphyaddr(struct kvm_vcpu *vcpu) kvm_vcpu 198 arch/x86/kvm/cpuid.c int kvm_vcpu_ioctl_set_cpuid(struct kvm_vcpu *vcpu, kvm_vcpu 243 arch/x86/kvm/cpuid.c int kvm_vcpu_ioctl_set_cpuid2(struct kvm_vcpu *vcpu, kvm_vcpu 264 arch/x86/kvm/cpuid.c int kvm_vcpu_ioctl_get_cpuid2(struct kvm_vcpu *vcpu, kvm_vcpu 928 arch/x86/kvm/cpuid.c static int move_to_next_stateful_cpuid_entry(struct kvm_vcpu *vcpu, int i) kvm_vcpu 962 arch/x86/kvm/cpuid.c struct kvm_cpuid_entry2 *kvm_find_cpuid_entry(struct kvm_vcpu *vcpu, kvm_vcpu 988 arch/x86/kvm/cpuid.c static bool cpuid_function_in_range(struct kvm_vcpu *vcpu, u32 function) kvm_vcpu 996 arch/x86/kvm/cpuid.c bool kvm_cpuid(struct kvm_vcpu *vcpu, u32 *eax, u32 *ebx, kvm_vcpu 1047 arch/x86/kvm/cpuid.c int kvm_emulate_cpuid(struct kvm_vcpu *vcpu) kvm_vcpu 9 arch/x86/kvm/cpuid.h int kvm_update_cpuid(struct kvm_vcpu *vcpu); kvm_vcpu 11 arch/x86/kvm/cpuid.h struct kvm_cpuid_entry2 *kvm_find_cpuid_entry(struct kvm_vcpu *vcpu, kvm_vcpu 16 arch/x86/kvm/cpuid.h int kvm_vcpu_ioctl_set_cpuid(struct kvm_vcpu *vcpu, kvm_vcpu 19 arch/x86/kvm/cpuid.h int kvm_vcpu_ioctl_set_cpuid2(struct kvm_vcpu *vcpu, kvm_vcpu 22 arch/x86/kvm/cpuid.h int kvm_vcpu_ioctl_get_cpuid2(struct kvm_vcpu *vcpu, kvm_vcpu 25 arch/x86/kvm/cpuid.h bool kvm_cpuid(struct kvm_vcpu *vcpu, u32 *eax, u32 *ebx, kvm_vcpu 28 arch/x86/kvm/cpuid.h int cpuid_query_maxphyaddr(struct kvm_vcpu *vcpu); kvm_vcpu 30 arch/x86/kvm/cpuid.h static inline int cpuid_maxphyaddr(struct kvm_vcpu *vcpu) kvm_vcpu 68 arch/x86/kvm/cpuid.h static __always_inline int *guest_cpuid_get_register(struct kvm_vcpu *vcpu, unsigned x86_feature) kvm_vcpu 92 arch/x86/kvm/cpuid.h static __always_inline bool guest_cpuid_has(struct kvm_vcpu *vcpu, unsigned x86_feature) kvm_vcpu 107 arch/x86/kvm/cpuid.h static __always_inline void guest_cpuid_clear(struct kvm_vcpu *vcpu, unsigned x86_feature) kvm_vcpu 116 arch/x86/kvm/cpuid.h static inline bool guest_cpuid_is_amd(struct kvm_vcpu *vcpu) kvm_vcpu 124 arch/x86/kvm/cpuid.h static inline int guest_cpuid_family(struct kvm_vcpu *vcpu) kvm_vcpu 135 arch/x86/kvm/cpuid.h static inline int guest_cpuid_model(struct kvm_vcpu *vcpu) kvm_vcpu 146 arch/x86/kvm/cpuid.h static inline int guest_cpuid_stepping(struct kvm_vcpu *vcpu) kvm_vcpu 157 arch/x86/kvm/cpuid.h static inline bool supports_cpuid_fault(struct kvm_vcpu *vcpu) kvm_vcpu 162 arch/x86/kvm/cpuid.h static inline bool cpuid_fault_enabled(struct kvm_vcpu *vcpu) kvm_vcpu 13 arch/x86/kvm/debugfs.c struct kvm_vcpu *vcpu = (struct kvm_vcpu *) data; kvm_vcpu 22 arch/x86/kvm/debugfs.c struct kvm_vcpu *vcpu = (struct kvm_vcpu *) data; kvm_vcpu 31 arch/x86/kvm/debugfs.c struct kvm_vcpu *vcpu = (struct kvm_vcpu *) data; kvm_vcpu 46 arch/x86/kvm/debugfs.c void kvm_arch_create_vcpu_debugfs(struct kvm_vcpu *vcpu) kvm_vcpu 133 arch/x86/kvm/hyperv.c static struct kvm_vcpu *get_vcpu_by_vpidx(struct kvm *kvm, u32 vpidx) kvm_vcpu 135 arch/x86/kvm/hyperv.c struct kvm_vcpu *vcpu = NULL; kvm_vcpu 152 arch/x86/kvm/hyperv.c struct kvm_vcpu *vcpu; kvm_vcpu 162 arch/x86/kvm/hyperv.c static void kvm_hv_notify_acked_sint(struct kvm_vcpu *vcpu, u32 sint) kvm_vcpu 190 arch/x86/kvm/hyperv.c struct kvm_vcpu *vcpu = synic_to_vcpu(synic); kvm_vcpu 205 arch/x86/kvm/hyperv.c struct kvm_vcpu *vcpu = synic_to_vcpu(synic); kvm_vcpu 305 arch/x86/kvm/hyperv.c struct kvm_vcpu *vcpu = synic_to_vcpu(synic); kvm_vcpu 339 arch/x86/kvm/hyperv.c void kvm_hv_synic_send_eoi(struct kvm_vcpu *vcpu, int vector) kvm_vcpu 399 arch/x86/kvm/hyperv.c struct kvm_vcpu *vcpu; kvm_vcpu 418 arch/x86/kvm/hyperv.c struct kvm_vcpu *vcpu = stimer_to_vcpu(stimer); kvm_vcpu 429 arch/x86/kvm/hyperv.c struct kvm_vcpu *vcpu = stimer_to_vcpu(stimer); kvm_vcpu 567 arch/x86/kvm/hyperv.c struct kvm_vcpu *vcpu = synic_to_vcpu(synic); kvm_vcpu 623 arch/x86/kvm/hyperv.c struct kvm_vcpu *vcpu = stimer_to_vcpu(stimer); kvm_vcpu 643 arch/x86/kvm/hyperv.c struct kvm_vcpu *vcpu = stimer_to_vcpu(stimer); kvm_vcpu 672 arch/x86/kvm/hyperv.c void kvm_hv_process_stimers(struct kvm_vcpu *vcpu) kvm_vcpu 702 arch/x86/kvm/hyperv.c void kvm_hv_vcpu_uninit(struct kvm_vcpu *vcpu) kvm_vcpu 711 arch/x86/kvm/hyperv.c bool kvm_hv_assist_page_enabled(struct kvm_vcpu *vcpu) kvm_vcpu 719 arch/x86/kvm/hyperv.c bool kvm_hv_get_assist_page(struct kvm_vcpu *vcpu, kvm_vcpu 753 arch/x86/kvm/hyperv.c void kvm_hv_vcpu_init(struct kvm_vcpu *vcpu) kvm_vcpu 765 arch/x86/kvm/hyperv.c void kvm_hv_vcpu_postcreate(struct kvm_vcpu *vcpu) kvm_vcpu 772 arch/x86/kvm/hyperv.c int kvm_hv_activate_synic(struct kvm_vcpu *vcpu, bool dont_zero_synic_pages) kvm_vcpu 808 arch/x86/kvm/hyperv.c static int kvm_hv_msr_get_crash_data(struct kvm_vcpu *vcpu, kvm_vcpu 821 arch/x86/kvm/hyperv.c static int kvm_hv_msr_get_crash_ctl(struct kvm_vcpu *vcpu, u64 *pdata) kvm_vcpu 829 arch/x86/kvm/hyperv.c static int kvm_hv_msr_set_crash_ctl(struct kvm_vcpu *vcpu, u64 data, bool host) kvm_vcpu 852 arch/x86/kvm/hyperv.c static int kvm_hv_msr_set_crash_data(struct kvm_vcpu *vcpu, kvm_vcpu 994 arch/x86/kvm/hyperv.c static int kvm_hv_set_msr_pw(struct kvm_vcpu *vcpu, u32 msr, u64 data, kvm_vcpu 1080 arch/x86/kvm/hyperv.c static int kvm_hv_set_msr(struct kvm_vcpu *vcpu, u32 msr, u64 data, bool host) kvm_vcpu 1191 arch/x86/kvm/hyperv.c static int kvm_hv_get_msr_pw(struct kvm_vcpu *vcpu, u32 msr, u64 *pdata) kvm_vcpu 1237 arch/x86/kvm/hyperv.c static int kvm_hv_get_msr(struct kvm_vcpu *vcpu, u32 msr, u64 *pdata, kvm_vcpu 1298 arch/x86/kvm/hyperv.c int kvm_hv_set_msr_common(struct kvm_vcpu *vcpu, u32 msr, u64 data, bool host) kvm_vcpu 1311 arch/x86/kvm/hyperv.c int kvm_hv_get_msr_common(struct kvm_vcpu *vcpu, u32 msr, u64 *pdata, bool host) kvm_vcpu 1329 arch/x86/kvm/hyperv.c struct kvm_vcpu *vcpu; kvm_vcpu 1352 arch/x86/kvm/hyperv.c static u64 kvm_hv_flush_tlb(struct kvm_vcpu *current_vcpu, u64 ingpa, kvm_vcpu 1443 arch/x86/kvm/hyperv.c struct kvm_vcpu *vcpu; kvm_vcpu 1455 arch/x86/kvm/hyperv.c static u64 kvm_hv_send_ipi(struct kvm_vcpu *current_vcpu, u64 ingpa, u64 outgpa, kvm_vcpu 1534 arch/x86/kvm/hyperv.c static void kvm_hv_hypercall_set_result(struct kvm_vcpu *vcpu, u64 result) kvm_vcpu 1547 arch/x86/kvm/hyperv.c static int kvm_hv_hypercall_complete(struct kvm_vcpu *vcpu, u64 result) kvm_vcpu 1554 arch/x86/kvm/hyperv.c static int kvm_hv_hypercall_complete_userspace(struct kvm_vcpu *vcpu) kvm_vcpu 1559 arch/x86/kvm/hyperv.c static u16 kvm_hvcall_signal_event(struct kvm_vcpu *vcpu, bool fast, u64 param) kvm_vcpu 1598 arch/x86/kvm/hyperv.c int kvm_hv_hypercall(struct kvm_vcpu *vcpu) kvm_vcpu 1786 arch/x86/kvm/hyperv.c int kvm_vcpu_ioctl_get_hv_cpuid(struct kvm_vcpu *vcpu, struct kvm_cpuid2 *cpuid, kvm_vcpu 26 arch/x86/kvm/hyperv.h static inline struct kvm_vcpu_hv *vcpu_to_hv_vcpu(struct kvm_vcpu *vcpu) kvm_vcpu 31 arch/x86/kvm/hyperv.h static inline struct kvm_vcpu *hv_vcpu_to_vcpu(struct kvm_vcpu_hv *hv_vcpu) kvm_vcpu 36 arch/x86/kvm/hyperv.h return container_of(arch, struct kvm_vcpu, arch); kvm_vcpu 39 arch/x86/kvm/hyperv.h static inline struct kvm_vcpu_hv_synic *vcpu_to_synic(struct kvm_vcpu *vcpu) kvm_vcpu 44 arch/x86/kvm/hyperv.h static inline struct kvm_vcpu *synic_to_vcpu(struct kvm_vcpu_hv_synic *synic) kvm_vcpu 49 arch/x86/kvm/hyperv.h int kvm_hv_set_msr_common(struct kvm_vcpu *vcpu, u32 msr, u64 data, bool host); kvm_vcpu 50 arch/x86/kvm/hyperv.h int kvm_hv_get_msr_common(struct kvm_vcpu *vcpu, u32 msr, u64 *pdata, bool host); kvm_vcpu 53 arch/x86/kvm/hyperv.h int kvm_hv_hypercall(struct kvm_vcpu *vcpu); kvm_vcpu 57 arch/x86/kvm/hyperv.h void kvm_hv_synic_send_eoi(struct kvm_vcpu *vcpu, int vector); kvm_vcpu 58 arch/x86/kvm/hyperv.h int kvm_hv_activate_synic(struct kvm_vcpu *vcpu, bool dont_zero_synic_pages); kvm_vcpu 60 arch/x86/kvm/hyperv.h void kvm_hv_vcpu_init(struct kvm_vcpu *vcpu); kvm_vcpu 61 arch/x86/kvm/hyperv.h void kvm_hv_vcpu_postcreate(struct kvm_vcpu *vcpu); kvm_vcpu 62 arch/x86/kvm/hyperv.h void kvm_hv_vcpu_uninit(struct kvm_vcpu *vcpu); kvm_vcpu 64 arch/x86/kvm/hyperv.h bool kvm_hv_assist_page_enabled(struct kvm_vcpu *vcpu); kvm_vcpu 65 arch/x86/kvm/hyperv.h bool kvm_hv_get_assist_page(struct kvm_vcpu *vcpu, kvm_vcpu 68 arch/x86/kvm/hyperv.h static inline struct kvm_vcpu_hv_stimer *vcpu_to_stimer(struct kvm_vcpu *vcpu, kvm_vcpu 74 arch/x86/kvm/hyperv.h static inline struct kvm_vcpu *stimer_to_vcpu(struct kvm_vcpu_hv_stimer *stimer) kvm_vcpu 83 arch/x86/kvm/hyperv.h static inline bool kvm_hv_has_stimer_pending(struct kvm_vcpu *vcpu) kvm_vcpu 89 arch/x86/kvm/hyperv.h void kvm_hv_process_stimers(struct kvm_vcpu *vcpu); kvm_vcpu 97 arch/x86/kvm/hyperv.h int kvm_vcpu_ioctl_get_hv_cpuid(struct kvm_vcpu *vcpu, struct kvm_cpuid2 *cpuid, kvm_vcpu 218 arch/x86/kvm/i8254.c void __kvm_migrate_pit_timer(struct kvm_vcpu *vcpu) kvm_vcpu 243 arch/x86/kvm/i8254.c struct kvm_vcpu *vcpu; kvm_vcpu 427 arch/x86/kvm/i8254.c static int pit_ioport_write(struct kvm_vcpu *vcpu, kvm_vcpu 503 arch/x86/kvm/i8254.c static int pit_ioport_read(struct kvm_vcpu *vcpu, kvm_vcpu 573 arch/x86/kvm/i8254.c static int speaker_ioport_write(struct kvm_vcpu *vcpu, kvm_vcpu 590 arch/x86/kvm/i8254.c static int speaker_ioport_read(struct kvm_vcpu *vcpu, kvm_vcpu 52 arch/x86/kvm/i8259.c struct kvm_vcpu *vcpu; kvm_vcpu 274 arch/x86/kvm/i8259.c struct kvm_vcpu *vcpu; kvm_vcpu 516 arch/x86/kvm/i8259.c static int picdev_master_write(struct kvm_vcpu *vcpu, struct kvm_io_device *dev, kvm_vcpu 523 arch/x86/kvm/i8259.c static int picdev_master_read(struct kvm_vcpu *vcpu, struct kvm_io_device *dev, kvm_vcpu 530 arch/x86/kvm/i8259.c static int picdev_slave_write(struct kvm_vcpu *vcpu, struct kvm_io_device *dev, kvm_vcpu 537 arch/x86/kvm/i8259.c static int picdev_slave_read(struct kvm_vcpu *vcpu, struct kvm_io_device *dev, kvm_vcpu 544 arch/x86/kvm/i8259.c static int picdev_eclr_write(struct kvm_vcpu *vcpu, struct kvm_io_device *dev, kvm_vcpu 551 arch/x86/kvm/i8259.c static int picdev_eclr_read(struct kvm_vcpu *vcpu, struct kvm_io_device *dev, kvm_vcpu 105 arch/x86/kvm/ioapic.c static void __rtc_irq_eoi_tracking_restore_one(struct kvm_vcpu *vcpu) kvm_vcpu 134 arch/x86/kvm/ioapic.c void kvm_rtc_eoi_tracking_restore_one(struct kvm_vcpu *vcpu) kvm_vcpu 145 arch/x86/kvm/ioapic.c struct kvm_vcpu *vcpu; kvm_vcpu 156 arch/x86/kvm/ioapic.c static void rtc_irq_eoi(struct kvm_ioapic *ioapic, struct kvm_vcpu *vcpu) kvm_vcpu 236 arch/x86/kvm/ioapic.c void kvm_ioapic_scan_entry(struct kvm_vcpu *vcpu, ulong *ioapic_handled_vectors) kvm_vcpu 422 arch/x86/kvm/ioapic.c static void __kvm_ioapic_update_eoi(struct kvm_vcpu *vcpu, kvm_vcpu 481 arch/x86/kvm/ioapic.c void kvm_ioapic_update_eoi(struct kvm_vcpu *vcpu, int vector, int trigger_mode) kvm_vcpu 501 arch/x86/kvm/ioapic.c static int ioapic_mmio_read(struct kvm_vcpu *vcpu, struct kvm_io_device *this, kvm_vcpu 543 arch/x86/kvm/ioapic.c static int ioapic_mmio_write(struct kvm_vcpu *vcpu, struct kvm_io_device *this, kvm_vcpu 10 arch/x86/kvm/ioapic.h struct kvm_vcpu; kvm_vcpu 118 arch/x86/kvm/ioapic.h void kvm_rtc_eoi_tracking_restore_one(struct kvm_vcpu *vcpu); kvm_vcpu 119 arch/x86/kvm/ioapic.h bool kvm_apic_match_dest(struct kvm_vcpu *vcpu, struct kvm_lapic *source, kvm_vcpu 121 arch/x86/kvm/ioapic.h int kvm_apic_compare_prio(struct kvm_vcpu *vcpu1, struct kvm_vcpu *vcpu2); kvm_vcpu 122 arch/x86/kvm/ioapic.h void kvm_ioapic_update_eoi(struct kvm_vcpu *vcpu, int vector, kvm_vcpu 134 arch/x86/kvm/ioapic.h void kvm_ioapic_scan_entry(struct kvm_vcpu *vcpu, kvm_vcpu 136 arch/x86/kvm/ioapic.h void kvm_scan_ioapic_routes(struct kvm_vcpu *vcpu, kvm_vcpu 22 arch/x86/kvm/irq.c int kvm_cpu_has_pending_timer(struct kvm_vcpu *vcpu) kvm_vcpu 34 arch/x86/kvm/irq.c static int pending_userspace_extint(struct kvm_vcpu *v) kvm_vcpu 43 arch/x86/kvm/irq.c static int kvm_cpu_has_extint(struct kvm_vcpu *v) kvm_vcpu 62 arch/x86/kvm/irq.c int kvm_cpu_has_injectable_intr(struct kvm_vcpu *v) kvm_vcpu 91 arch/x86/kvm/irq.c int kvm_cpu_has_interrupt(struct kvm_vcpu *v) kvm_vcpu 118 arch/x86/kvm/irq.c static int kvm_cpu_get_extint(struct kvm_vcpu *v) kvm_vcpu 135 arch/x86/kvm/irq.c int kvm_cpu_get_interrupt(struct kvm_vcpu *v) kvm_vcpu 151 arch/x86/kvm/irq.c void kvm_inject_pending_timer_irqs(struct kvm_vcpu *vcpu) kvm_vcpu 158 arch/x86/kvm/irq.c void __kvm_migrate_timers(struct kvm_vcpu *vcpu) kvm_vcpu 27 arch/x86/kvm/irq.h struct kvm_vcpu; kvm_vcpu 105 arch/x86/kvm/irq.h void kvm_inject_pending_timer_irqs(struct kvm_vcpu *vcpu); kvm_vcpu 106 arch/x86/kvm/irq.h void kvm_inject_apic_timer_irqs(struct kvm_vcpu *vcpu); kvm_vcpu 107 arch/x86/kvm/irq.h void kvm_apic_nmi_wd_deliver(struct kvm_vcpu *vcpu); kvm_vcpu 108 arch/x86/kvm/irq.h void __kvm_migrate_apic_timer(struct kvm_vcpu *vcpu); kvm_vcpu 109 arch/x86/kvm/irq.h void __kvm_migrate_pit_timer(struct kvm_vcpu *vcpu); kvm_vcpu 110 arch/x86/kvm/irq.h void __kvm_migrate_timers(struct kvm_vcpu *vcpu); kvm_vcpu 112 arch/x86/kvm/irq.h int apic_has_pending_timer(struct kvm_vcpu *vcpu); kvm_vcpu 51 arch/x86/kvm/irq_comm.c struct kvm_vcpu *vcpu, *lowest = NULL; kvm_vcpu 325 arch/x86/kvm/irq_comm.c struct kvm_vcpu **dest_vcpu) kvm_vcpu 328 arch/x86/kvm/irq_comm.c struct kvm_vcpu *vcpu; kvm_vcpu 397 arch/x86/kvm/irq_comm.c void kvm_scan_ioapic_routes(struct kvm_vcpu *vcpu, kvm_vcpu 13 arch/x86/kvm/kvm_cache_regs.h static __always_inline unsigned long kvm_##lname##_read(struct kvm_vcpu *vcpu)\ kvm_vcpu 17 arch/x86/kvm/kvm_cache_regs.h static __always_inline void kvm_##lname##_write(struct kvm_vcpu *vcpu, \ kvm_vcpu 40 arch/x86/kvm/kvm_cache_regs.h static inline unsigned long kvm_register_read(struct kvm_vcpu *vcpu, kvm_vcpu 49 arch/x86/kvm/kvm_cache_regs.h static inline void kvm_register_write(struct kvm_vcpu *vcpu, kvm_vcpu 58 arch/x86/kvm/kvm_cache_regs.h static inline unsigned long kvm_rip_read(struct kvm_vcpu *vcpu) kvm_vcpu 63 arch/x86/kvm/kvm_cache_regs.h static inline void kvm_rip_write(struct kvm_vcpu *vcpu, unsigned long val) kvm_vcpu 68 arch/x86/kvm/kvm_cache_regs.h static inline unsigned long kvm_rsp_read(struct kvm_vcpu *vcpu) kvm_vcpu 73 arch/x86/kvm/kvm_cache_regs.h static inline void kvm_rsp_write(struct kvm_vcpu *vcpu, unsigned long val) kvm_vcpu 78 arch/x86/kvm/kvm_cache_regs.h static inline u64 kvm_pdptr_read(struct kvm_vcpu *vcpu, int index) kvm_vcpu 89 arch/x86/kvm/kvm_cache_regs.h static inline ulong kvm_read_cr0_bits(struct kvm_vcpu *vcpu, ulong mask) kvm_vcpu 97 arch/x86/kvm/kvm_cache_regs.h static inline ulong kvm_read_cr0(struct kvm_vcpu *vcpu) kvm_vcpu 102 arch/x86/kvm/kvm_cache_regs.h static inline ulong kvm_read_cr4_bits(struct kvm_vcpu *vcpu, ulong mask) kvm_vcpu 110 arch/x86/kvm/kvm_cache_regs.h static inline ulong kvm_read_cr3(struct kvm_vcpu *vcpu) kvm_vcpu 117 arch/x86/kvm/kvm_cache_regs.h static inline ulong kvm_read_cr4(struct kvm_vcpu *vcpu) kvm_vcpu 122 arch/x86/kvm/kvm_cache_regs.h static inline u64 kvm_read_edx_eax(struct kvm_vcpu *vcpu) kvm_vcpu 128 arch/x86/kvm/kvm_cache_regs.h static inline void enter_guest_mode(struct kvm_vcpu *vcpu) kvm_vcpu 133 arch/x86/kvm/kvm_cache_regs.h static inline void leave_guest_mode(struct kvm_vcpu *vcpu) kvm_vcpu 143 arch/x86/kvm/kvm_cache_regs.h static inline bool is_guest_mode(struct kvm_vcpu *vcpu) kvm_vcpu 148 arch/x86/kvm/kvm_cache_regs.h static inline bool is_smm(struct kvm_vcpu *vcpu) kvm_vcpu 81 arch/x86/kvm/lapic.c bool kvm_apic_pending_eoi(struct kvm_vcpu *vcpu, int vector) kvm_vcpu 119 arch/x86/kvm/lapic.c bool kvm_can_post_timer_interrupt(struct kvm_vcpu *vcpu) kvm_vcpu 125 arch/x86/kvm/lapic.c static bool kvm_use_posted_timer_interrupt(struct kvm_vcpu *vcpu) kvm_vcpu 173 arch/x86/kvm/lapic.c struct kvm_vcpu *vcpu; kvm_vcpu 325 arch/x86/kvm/lapic.c void kvm_apic_set_version(struct kvm_vcpu *vcpu) kvm_vcpu 415 arch/x86/kvm/lapic.c bool kvm_apic_update_irr(struct kvm_vcpu *vcpu, u32 *pir, int *max_irr) kvm_vcpu 447 arch/x86/kvm/lapic.c struct kvm_vcpu *vcpu; kvm_vcpu 466 arch/x86/kvm/lapic.c struct kvm_vcpu *vcpu; kvm_vcpu 513 arch/x86/kvm/lapic.c struct kvm_vcpu *vcpu; kvm_vcpu 536 arch/x86/kvm/lapic.c int kvm_lapic_find_highest_irr(struct kvm_vcpu *vcpu) kvm_vcpu 551 arch/x86/kvm/lapic.c int kvm_apic_set_irq(struct kvm_vcpu *vcpu, struct kvm_lapic_irq *irq, kvm_vcpu 566 arch/x86/kvm/lapic.c struct kvm_vcpu *vcpu; kvm_vcpu 618 arch/x86/kvm/lapic.c static int pv_eoi_put_user(struct kvm_vcpu *vcpu, u8 val) kvm_vcpu 625 arch/x86/kvm/lapic.c static int pv_eoi_get_user(struct kvm_vcpu *vcpu, u8 *val) kvm_vcpu 632 arch/x86/kvm/lapic.c static inline bool pv_eoi_enabled(struct kvm_vcpu *vcpu) kvm_vcpu 637 arch/x86/kvm/lapic.c static bool pv_eoi_get_pending(struct kvm_vcpu *vcpu) kvm_vcpu 648 arch/x86/kvm/lapic.c static void pv_eoi_set_pending(struct kvm_vcpu *vcpu) kvm_vcpu 658 arch/x86/kvm/lapic.c static void pv_eoi_clr_pending(struct kvm_vcpu *vcpu) kvm_vcpu 711 arch/x86/kvm/lapic.c void kvm_apic_update_ppr(struct kvm_vcpu *vcpu) kvm_vcpu 791 arch/x86/kvm/lapic.c static u32 kvm_apic_mda(struct kvm_vcpu *vcpu, unsigned int dest_id, kvm_vcpu 803 arch/x86/kvm/lapic.c bool kvm_apic_match_dest(struct kvm_vcpu *vcpu, struct kvm_lapic *source, kvm_vcpu 993 arch/x86/kvm/lapic.c struct kvm_vcpu **dest_vcpu) kvm_vcpu 1029 arch/x86/kvm/lapic.c struct kvm_vcpu *vcpu = apic->vcpu; kvm_vcpu 1126 arch/x86/kvm/lapic.c int kvm_apic_compare_prio(struct kvm_vcpu *vcpu1, struct kvm_vcpu *vcpu2) kvm_vcpu 1187 arch/x86/kvm/lapic.c void kvm_apic_set_eoi_accelerated(struct kvm_vcpu *vcpu, int vector) kvm_vcpu 1246 arch/x86/kvm/lapic.c struct kvm_vcpu *vcpu = apic->vcpu; kvm_vcpu 1363 arch/x86/kvm/lapic.c static int apic_mmio_read(struct kvm_vcpu *vcpu, struct kvm_io_device *this, kvm_vcpu 1440 arch/x86/kvm/lapic.c static bool lapic_timer_int_injected(struct kvm_vcpu *vcpu) kvm_vcpu 1458 arch/x86/kvm/lapic.c static inline void __wait_lapic_expire(struct kvm_vcpu *vcpu, u64 guest_cycles) kvm_vcpu 1478 arch/x86/kvm/lapic.c static inline void adjust_lapic_timer_advance(struct kvm_vcpu *vcpu, kvm_vcpu 1507 arch/x86/kvm/lapic.c static void __kvm_wait_lapic_expire(struct kvm_vcpu *vcpu) kvm_vcpu 1527 arch/x86/kvm/lapic.c void kvm_wait_lapic_expire(struct kvm_vcpu *vcpu) kvm_vcpu 1549 arch/x86/kvm/lapic.c struct kvm_vcpu *vcpu = apic->vcpu; kvm_vcpu 1575 arch/x86/kvm/lapic.c struct kvm_vcpu *vcpu = apic->vcpu; kvm_vcpu 1690 arch/x86/kvm/lapic.c bool kvm_lapic_hv_timer_in_use(struct kvm_vcpu *vcpu) kvm_vcpu 1710 arch/x86/kvm/lapic.c struct kvm_vcpu *vcpu = apic->vcpu; kvm_vcpu 1779 arch/x86/kvm/lapic.c void kvm_lapic_expired_hv_timer(struct kvm_vcpu *vcpu) kvm_vcpu 1800 arch/x86/kvm/lapic.c void kvm_lapic_switch_to_hv_timer(struct kvm_vcpu *vcpu) kvm_vcpu 1806 arch/x86/kvm/lapic.c void kvm_lapic_switch_to_sw_timer(struct kvm_vcpu *vcpu) kvm_vcpu 1818 arch/x86/kvm/lapic.c void kvm_lapic_restart_hv_timer(struct kvm_vcpu *vcpu) kvm_vcpu 1993 arch/x86/kvm/lapic.c static int apic_mmio_write(struct kvm_vcpu *vcpu, struct kvm_io_device *this, kvm_vcpu 2026 arch/x86/kvm/lapic.c void kvm_lapic_set_eoi(struct kvm_vcpu *vcpu) kvm_vcpu 2033 arch/x86/kvm/lapic.c void kvm_apic_write_nodecode(struct kvm_vcpu *vcpu, u32 offset) kvm_vcpu 2047 arch/x86/kvm/lapic.c void kvm_free_lapic(struct kvm_vcpu *vcpu) kvm_vcpu 2073 arch/x86/kvm/lapic.c u64 kvm_get_lapic_tscdeadline_msr(struct kvm_vcpu *vcpu) kvm_vcpu 2084 arch/x86/kvm/lapic.c void kvm_set_lapic_tscdeadline_msr(struct kvm_vcpu *vcpu, u64 data) kvm_vcpu 2097 arch/x86/kvm/lapic.c void kvm_lapic_set_tpr(struct kvm_vcpu *vcpu, unsigned long cr8) kvm_vcpu 2105 arch/x86/kvm/lapic.c u64 kvm_lapic_get_cr8(struct kvm_vcpu *vcpu) kvm_vcpu 2114 arch/x86/kvm/lapic.c void kvm_lapic_set_base(struct kvm_vcpu *vcpu, u64 value) kvm_vcpu 2155 arch/x86/kvm/lapic.c void kvm_lapic_reset(struct kvm_vcpu *vcpu, bool init_event) kvm_vcpu 2228 arch/x86/kvm/lapic.c int apic_has_pending_timer(struct kvm_vcpu *vcpu) kvm_vcpu 2253 arch/x86/kvm/lapic.c void kvm_apic_nmi_wd_deliver(struct kvm_vcpu *vcpu) kvm_vcpu 2281 arch/x86/kvm/lapic.c int kvm_create_lapic(struct kvm_vcpu *vcpu, int timer_advance_ns) kvm_vcpu 2328 arch/x86/kvm/lapic.c int kvm_apic_has_interrupt(struct kvm_vcpu *vcpu) kvm_vcpu 2340 arch/x86/kvm/lapic.c int kvm_apic_accept_pic_intr(struct kvm_vcpu *vcpu) kvm_vcpu 2353 arch/x86/kvm/lapic.c void kvm_inject_apic_timer_irqs(struct kvm_vcpu *vcpu) kvm_vcpu 2363 arch/x86/kvm/lapic.c int kvm_get_apic_interrupt(struct kvm_vcpu *vcpu) kvm_vcpu 2401 arch/x86/kvm/lapic.c static int kvm_apic_state_fixup(struct kvm_vcpu *vcpu, kvm_vcpu 2426 arch/x86/kvm/lapic.c int kvm_apic_get_state(struct kvm_vcpu *vcpu, struct kvm_lapic_state *s) kvm_vcpu 2432 arch/x86/kvm/lapic.c int kvm_apic_set_state(struct kvm_vcpu *vcpu, struct kvm_lapic_state *s) kvm_vcpu 2476 arch/x86/kvm/lapic.c void __kvm_migrate_apic_timer(struct kvm_vcpu *vcpu) kvm_vcpu 2496 arch/x86/kvm/lapic.c static void apic_sync_pv_eoi_from_guest(struct kvm_vcpu *vcpu, kvm_vcpu 2526 arch/x86/kvm/lapic.c void kvm_lapic_sync_from_vapic(struct kvm_vcpu *vcpu) kvm_vcpu 2549 arch/x86/kvm/lapic.c static void apic_sync_pv_eoi_to_guest(struct kvm_vcpu *vcpu, kvm_vcpu 2569 arch/x86/kvm/lapic.c void kvm_lapic_sync_to_vapic(struct kvm_vcpu *vcpu) kvm_vcpu 2593 arch/x86/kvm/lapic.c int kvm_lapic_set_vapic_addr(struct kvm_vcpu *vcpu, gpa_t vapic_addr) kvm_vcpu 2609 arch/x86/kvm/lapic.c int kvm_x2apic_msr_write(struct kvm_vcpu *vcpu, u32 msr, u64 data) kvm_vcpu 2626 arch/x86/kvm/lapic.c int kvm_x2apic_msr_read(struct kvm_vcpu *vcpu, u32 msr, u64 *data) kvm_vcpu 2647 arch/x86/kvm/lapic.c int kvm_hv_vapic_msr_write(struct kvm_vcpu *vcpu, u32 reg, u64 data) kvm_vcpu 2660 arch/x86/kvm/lapic.c int kvm_hv_vapic_msr_read(struct kvm_vcpu *vcpu, u32 reg, u64 *data) kvm_vcpu 2678 arch/x86/kvm/lapic.c int kvm_lapic_enable_pv_eoi(struct kvm_vcpu *vcpu, u64 data, unsigned long len) kvm_vcpu 2699 arch/x86/kvm/lapic.c void kvm_apic_accept_events(struct kvm_vcpu *vcpu) kvm_vcpu 45 arch/x86/kvm/lapic.h struct kvm_vcpu *vcpu; kvm_vcpu 67 arch/x86/kvm/lapic.h int kvm_create_lapic(struct kvm_vcpu *vcpu, int timer_advance_ns); kvm_vcpu 68 arch/x86/kvm/lapic.h void kvm_free_lapic(struct kvm_vcpu *vcpu); kvm_vcpu 70 arch/x86/kvm/lapic.h int kvm_apic_has_interrupt(struct kvm_vcpu *vcpu); kvm_vcpu 71 arch/x86/kvm/lapic.h int kvm_apic_accept_pic_intr(struct kvm_vcpu *vcpu); kvm_vcpu 72 arch/x86/kvm/lapic.h int kvm_get_apic_interrupt(struct kvm_vcpu *vcpu); kvm_vcpu 73 arch/x86/kvm/lapic.h void kvm_apic_accept_events(struct kvm_vcpu *vcpu); kvm_vcpu 74 arch/x86/kvm/lapic.h void kvm_lapic_reset(struct kvm_vcpu *vcpu, bool init_event); kvm_vcpu 75 arch/x86/kvm/lapic.h u64 kvm_lapic_get_cr8(struct kvm_vcpu *vcpu); kvm_vcpu 76 arch/x86/kvm/lapic.h void kvm_lapic_set_tpr(struct kvm_vcpu *vcpu, unsigned long cr8); kvm_vcpu 77 arch/x86/kvm/lapic.h void kvm_lapic_set_eoi(struct kvm_vcpu *vcpu); kvm_vcpu 78 arch/x86/kvm/lapic.h void kvm_lapic_set_base(struct kvm_vcpu *vcpu, u64 value); kvm_vcpu 79 arch/x86/kvm/lapic.h u64 kvm_lapic_get_base(struct kvm_vcpu *vcpu); kvm_vcpu 80 arch/x86/kvm/lapic.h void kvm_apic_set_version(struct kvm_vcpu *vcpu); kvm_vcpu 84 arch/x86/kvm/lapic.h bool kvm_apic_match_dest(struct kvm_vcpu *vcpu, struct kvm_lapic *source, kvm_vcpu 88 arch/x86/kvm/lapic.h bool kvm_apic_update_irr(struct kvm_vcpu *vcpu, u32 *pir, int *max_irr); kvm_vcpu 89 arch/x86/kvm/lapic.h void kvm_apic_update_ppr(struct kvm_vcpu *vcpu); kvm_vcpu 90 arch/x86/kvm/lapic.h int kvm_apic_set_irq(struct kvm_vcpu *vcpu, struct kvm_lapic_irq *irq, kvm_vcpu 97 arch/x86/kvm/lapic.h u64 kvm_get_apic_base(struct kvm_vcpu *vcpu); kvm_vcpu 98 arch/x86/kvm/lapic.h int kvm_set_apic_base(struct kvm_vcpu *vcpu, struct msr_data *msr_info); kvm_vcpu 99 arch/x86/kvm/lapic.h int kvm_apic_get_state(struct kvm_vcpu *vcpu, struct kvm_lapic_state *s); kvm_vcpu 100 arch/x86/kvm/lapic.h int kvm_apic_set_state(struct kvm_vcpu *vcpu, struct kvm_lapic_state *s); kvm_vcpu 101 arch/x86/kvm/lapic.h enum lapic_mode kvm_get_apic_mode(struct kvm_vcpu *vcpu); kvm_vcpu 102 arch/x86/kvm/lapic.h int kvm_lapic_find_highest_irr(struct kvm_vcpu *vcpu); kvm_vcpu 104 arch/x86/kvm/lapic.h u64 kvm_get_lapic_tscdeadline_msr(struct kvm_vcpu *vcpu); kvm_vcpu 105 arch/x86/kvm/lapic.h void kvm_set_lapic_tscdeadline_msr(struct kvm_vcpu *vcpu, u64 data); kvm_vcpu 107 arch/x86/kvm/lapic.h void kvm_apic_write_nodecode(struct kvm_vcpu *vcpu, u32 offset); kvm_vcpu 108 arch/x86/kvm/lapic.h void kvm_apic_set_eoi_accelerated(struct kvm_vcpu *vcpu, int vector); kvm_vcpu 110 arch/x86/kvm/lapic.h int kvm_lapic_set_vapic_addr(struct kvm_vcpu *vcpu, gpa_t vapic_addr); kvm_vcpu 111 arch/x86/kvm/lapic.h void kvm_lapic_sync_from_vapic(struct kvm_vcpu *vcpu); kvm_vcpu 112 arch/x86/kvm/lapic.h void kvm_lapic_sync_to_vapic(struct kvm_vcpu *vcpu); kvm_vcpu 114 arch/x86/kvm/lapic.h int kvm_x2apic_msr_write(struct kvm_vcpu *vcpu, u32 msr, u64 data); kvm_vcpu 115 arch/x86/kvm/lapic.h int kvm_x2apic_msr_read(struct kvm_vcpu *vcpu, u32 msr, u64 *data); kvm_vcpu 117 arch/x86/kvm/lapic.h int kvm_hv_vapic_msr_write(struct kvm_vcpu *vcpu, u32 msr, u64 data); kvm_vcpu 118 arch/x86/kvm/lapic.h int kvm_hv_vapic_msr_read(struct kvm_vcpu *vcpu, u32 msr, u64 *data); kvm_vcpu 120 arch/x86/kvm/lapic.h static inline bool kvm_hv_vapic_assist_page_enabled(struct kvm_vcpu *vcpu) kvm_vcpu 125 arch/x86/kvm/lapic.h int kvm_lapic_enable_pv_eoi(struct kvm_vcpu *vcpu, u64 data, unsigned long len); kvm_vcpu 164 arch/x86/kvm/lapic.h static inline bool lapic_in_kernel(struct kvm_vcpu *vcpu) kvm_vcpu 189 arch/x86/kvm/lapic.h static inline bool kvm_apic_present(struct kvm_vcpu *vcpu) kvm_vcpu 194 arch/x86/kvm/lapic.h static inline int kvm_lapic_enabled(struct kvm_vcpu *vcpu) kvm_vcpu 204 arch/x86/kvm/lapic.h static inline bool kvm_vcpu_apicv_active(struct kvm_vcpu *vcpu) kvm_vcpu 209 arch/x86/kvm/lapic.h static inline bool kvm_apic_has_events(struct kvm_vcpu *vcpu) kvm_vcpu 220 arch/x86/kvm/lapic.h static inline int kvm_lapic_latched_init(struct kvm_vcpu *vcpu) kvm_vcpu 225 arch/x86/kvm/lapic.h bool kvm_apic_pending_eoi(struct kvm_vcpu *vcpu, int vector); kvm_vcpu 227 arch/x86/kvm/lapic.h void kvm_wait_lapic_expire(struct kvm_vcpu *vcpu); kvm_vcpu 230 arch/x86/kvm/lapic.h struct kvm_vcpu **dest_vcpu); kvm_vcpu 233 arch/x86/kvm/lapic.h void kvm_lapic_switch_to_sw_timer(struct kvm_vcpu *vcpu); kvm_vcpu 234 arch/x86/kvm/lapic.h void kvm_lapic_switch_to_hv_timer(struct kvm_vcpu *vcpu); kvm_vcpu 235 arch/x86/kvm/lapic.h void kvm_lapic_expired_hv_timer(struct kvm_vcpu *vcpu); kvm_vcpu 236 arch/x86/kvm/lapic.h bool kvm_lapic_hv_timer_in_use(struct kvm_vcpu *vcpu); kvm_vcpu 237 arch/x86/kvm/lapic.h void kvm_lapic_restart_hv_timer(struct kvm_vcpu *vcpu); kvm_vcpu 238 arch/x86/kvm/lapic.h bool kvm_can_post_timer_interrupt(struct kvm_vcpu *vcpu); kvm_vcpu 308 arch/x86/kvm/mmu.c kvm_mmu_calc_root_page_role(struct kvm_vcpu *vcpu); kvm_vcpu 364 arch/x86/kvm/mmu.c static inline bool kvm_vcpu_ad_need_write_protect(struct kvm_vcpu *vcpu) kvm_vcpu 456 arch/x86/kvm/mmu.c static void mark_mmio_spte(struct kvm_vcpu *vcpu, u64 *sptep, u64 gfn, kvm_vcpu 488 arch/x86/kvm/mmu.c static bool set_mmio_spte(struct kvm_vcpu *vcpu, u64 *sptep, gfn_t gfn, kvm_vcpu 499 arch/x86/kvm/mmu.c static bool check_mmio_spte(struct kvm_vcpu *vcpu, u64 spte) kvm_vcpu 604 arch/x86/kvm/mmu.c static int is_nx(struct kvm_vcpu *vcpu) kvm_vcpu 1034 arch/x86/kvm/mmu.c static void walk_shadow_page_lockless_begin(struct kvm_vcpu *vcpu) kvm_vcpu 1049 arch/x86/kvm/mmu.c static void walk_shadow_page_lockless_end(struct kvm_vcpu *vcpu) kvm_vcpu 1110 arch/x86/kvm/mmu.c static int mmu_topup_memory_caches(struct kvm_vcpu *vcpu) kvm_vcpu 1127 arch/x86/kvm/mmu.c static void mmu_free_memory_caches(struct kvm_vcpu *vcpu) kvm_vcpu 1145 arch/x86/kvm/mmu.c static struct pte_list_desc *mmu_alloc_pte_list_desc(struct kvm_vcpu *vcpu) kvm_vcpu 1281 arch/x86/kvm/mmu.c static bool mmu_gfn_lpage_is_disallowed(struct kvm_vcpu *vcpu, gfn_t gfn, kvm_vcpu 1290 arch/x86/kvm/mmu.c static int host_mapping_level(struct kvm_vcpu *vcpu, gfn_t gfn) kvm_vcpu 1319 arch/x86/kvm/mmu.c gfn_to_memslot_dirty_bitmap(struct kvm_vcpu *vcpu, gfn_t gfn, kvm_vcpu 1331 arch/x86/kvm/mmu.c static int mapping_level(struct kvm_vcpu *vcpu, gfn_t large_gfn, kvm_vcpu 1370 arch/x86/kvm/mmu.c static int pte_list_add(struct kvm_vcpu *vcpu, u64 *spte, kvm_vcpu 1489 arch/x86/kvm/mmu.c static bool rmap_can_add(struct kvm_vcpu *vcpu) kvm_vcpu 1497 arch/x86/kvm/mmu.c static int rmap_add(struct kvm_vcpu *vcpu, u64 *spte, gfn_t gfn) kvm_vcpu 1616 arch/x86/kvm/mmu.c static void drop_large_spte(struct kvm_vcpu *vcpu, u64 *sptep) kvm_vcpu 1822 arch/x86/kvm/mmu.c int kvm_arch_write_log_dirty(struct kvm_vcpu *vcpu) kvm_vcpu 1845 arch/x86/kvm/mmu.c static bool rmap_write_protect(struct kvm_vcpu *vcpu, u64 gfn) kvm_vcpu 2088 arch/x86/kvm/mmu.c static void rmap_recycle(struct kvm_vcpu *vcpu, u64 *spte, gfn_t gfn) kvm_vcpu 2156 arch/x86/kvm/mmu.c static void mmu_page_add_parent_pte(struct kvm_vcpu *vcpu, kvm_vcpu 2178 arch/x86/kvm/mmu.c static struct kvm_mmu_page *kvm_mmu_alloc_page(struct kvm_vcpu *vcpu, int direct) kvm_vcpu 2224 arch/x86/kvm/mmu.c static int nonpaging_sync_page(struct kvm_vcpu *vcpu, kvm_vcpu 2230 arch/x86/kvm/mmu.c static void nonpaging_invlpg(struct kvm_vcpu *vcpu, gva_t gva, hpa_t root) kvm_vcpu 2234 arch/x86/kvm/mmu.c static void nonpaging_update_pte(struct kvm_vcpu *vcpu, kvm_vcpu 2356 arch/x86/kvm/mmu.c static bool __kvm_sync_page(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp, kvm_vcpu 2382 arch/x86/kvm/mmu.c static void kvm_mmu_flush_or_zap(struct kvm_vcpu *vcpu, kvm_vcpu 2396 arch/x86/kvm/mmu.c static void kvm_mmu_audit(struct kvm_vcpu *vcpu, int point) { } kvm_vcpu 2406 arch/x86/kvm/mmu.c static bool kvm_sync_page(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp, kvm_vcpu 2414 arch/x86/kvm/mmu.c static bool kvm_sync_pages(struct kvm_vcpu *vcpu, gfn_t gfn, kvm_vcpu 2503 arch/x86/kvm/mmu.c static void mmu_sync_children(struct kvm_vcpu *vcpu, kvm_vcpu 2550 arch/x86/kvm/mmu.c static struct kvm_mmu_page *kvm_mmu_get_page(struct kvm_vcpu *vcpu, kvm_vcpu 2641 arch/x86/kvm/mmu.c struct kvm_vcpu *vcpu, hpa_t root, kvm_vcpu 2670 arch/x86/kvm/mmu.c struct kvm_vcpu *vcpu, u64 addr) kvm_vcpu 2703 arch/x86/kvm/mmu.c static void link_shadow_page(struct kvm_vcpu *vcpu, u64 *sptep, kvm_vcpu 2726 arch/x86/kvm/mmu.c static void validate_direct_spte(struct kvm_vcpu *vcpu, u64 *sptep, kvm_vcpu 2954 arch/x86/kvm/mmu.c static void kvm_unsync_page(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp) kvm_vcpu 2963 arch/x86/kvm/mmu.c static bool mmu_need_write_protect(struct kvm_vcpu *vcpu, gfn_t gfn, kvm_vcpu 3049 arch/x86/kvm/mmu.c static int set_spte(struct kvm_vcpu *vcpu, u64 *sptep, kvm_vcpu 3153 arch/x86/kvm/mmu.c static int mmu_set_spte(struct kvm_vcpu *vcpu, u64 *sptep, unsigned pte_access, kvm_vcpu 3219 arch/x86/kvm/mmu.c static kvm_pfn_t pte_prefetch_gfn_to_pfn(struct kvm_vcpu *vcpu, gfn_t gfn, kvm_vcpu 3231 arch/x86/kvm/mmu.c static int direct_pte_prefetch_many(struct kvm_vcpu *vcpu, kvm_vcpu 3259 arch/x86/kvm/mmu.c static void __direct_pte_prefetch(struct kvm_vcpu *vcpu, kvm_vcpu 3282 arch/x86/kvm/mmu.c static void direct_pte_prefetch(struct kvm_vcpu *vcpu, u64 *sptep) kvm_vcpu 3325 arch/x86/kvm/mmu.c static int __direct_map(struct kvm_vcpu *vcpu, gpa_t gpa, int write, kvm_vcpu 3374 arch/x86/kvm/mmu.c static int kvm_handle_bad_page(struct kvm_vcpu *vcpu, gfn_t gfn, kvm_pfn_t pfn) kvm_vcpu 3392 arch/x86/kvm/mmu.c static void transparent_hugepage_adjust(struct kvm_vcpu *vcpu, kvm_vcpu 3431 arch/x86/kvm/mmu.c static bool handle_abnormal_pfn(struct kvm_vcpu *vcpu, gva_t gva, gfn_t gfn, kvm_vcpu 3485 arch/x86/kvm/mmu.c fast_pf_fix_direct_spte(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp, kvm_vcpu 3536 arch/x86/kvm/mmu.c static bool fast_page_fault(struct kvm_vcpu *vcpu, gpa_t cr2_or_gpa, int level, kvm_vcpu 3641 arch/x86/kvm/mmu.c static bool try_async_pf(struct kvm_vcpu *vcpu, bool prefault, gfn_t gfn, kvm_vcpu 3644 arch/x86/kvm/mmu.c static int make_mmu_pages_available(struct kvm_vcpu *vcpu); kvm_vcpu 3646 arch/x86/kvm/mmu.c static int nonpaging_map(struct kvm_vcpu *vcpu, gpa_t gpa, u32 error_code, kvm_vcpu 3717 arch/x86/kvm/mmu.c void kvm_mmu_free_roots(struct kvm_vcpu *vcpu, struct kvm_mmu *mmu, kvm_vcpu 3765 arch/x86/kvm/mmu.c static int mmu_check_root(struct kvm_vcpu *vcpu, gfn_t root_gfn) kvm_vcpu 3777 arch/x86/kvm/mmu.c static int mmu_alloc_direct_roots(struct kvm_vcpu *vcpu) kvm_vcpu 3818 arch/x86/kvm/mmu.c static int mmu_alloc_shadow_roots(struct kvm_vcpu *vcpu) kvm_vcpu 3923 arch/x86/kvm/mmu.c static int mmu_alloc_roots(struct kvm_vcpu *vcpu) kvm_vcpu 3931 arch/x86/kvm/mmu.c void kvm_mmu_sync_roots(struct kvm_vcpu *vcpu) kvm_vcpu 3990 arch/x86/kvm/mmu.c static gpa_t nonpaging_gva_to_gpa(struct kvm_vcpu *vcpu, gpa_t vaddr, kvm_vcpu 3998 arch/x86/kvm/mmu.c static gpa_t nonpaging_gva_to_gpa_nested(struct kvm_vcpu *vcpu, gpa_t vaddr, kvm_vcpu 4026 arch/x86/kvm/mmu.c static bool mmio_info_in_cache(struct kvm_vcpu *vcpu, u64 addr, bool direct) kvm_vcpu 4043 arch/x86/kvm/mmu.c walk_shadow_page_get_mmio_spte(struct kvm_vcpu *vcpu, u64 addr, u64 *sptep) kvm_vcpu 4087 arch/x86/kvm/mmu.c static int handle_mmio_page_fault(struct kvm_vcpu *vcpu, u64 addr, bool direct) kvm_vcpu 4121 arch/x86/kvm/mmu.c static bool page_fault_handle_page_track(struct kvm_vcpu *vcpu, kvm_vcpu 4141 arch/x86/kvm/mmu.c static void shadow_page_table_clear_flood(struct kvm_vcpu *vcpu, gva_t addr) kvm_vcpu 4158 arch/x86/kvm/mmu.c static int nonpaging_page_fault(struct kvm_vcpu *vcpu, gpa_t gpa, kvm_vcpu 4181 arch/x86/kvm/mmu.c static int kvm_arch_setup_async_pf(struct kvm_vcpu *vcpu, gpa_t cr2_or_gpa, kvm_vcpu 4195 arch/x86/kvm/mmu.c static bool try_async_pf(struct kvm_vcpu *vcpu, bool prefault, gfn_t gfn, kvm_vcpu 4230 arch/x86/kvm/mmu.c int kvm_handle_page_fault(struct kvm_vcpu *vcpu, u64 error_code, kvm_vcpu 4269 arch/x86/kvm/mmu.c check_hugepage_cache_consistency(struct kvm_vcpu *vcpu, gfn_t gfn, int level) kvm_vcpu 4278 arch/x86/kvm/mmu.c static int tdp_page_fault(struct kvm_vcpu *vcpu, gpa_t gpa, u32 error_code, kvm_vcpu 4340 arch/x86/kvm/mmu.c static void nonpaging_init_context(struct kvm_vcpu *vcpu, kvm_vcpu 4362 arch/x86/kvm/mmu.c static bool cached_root_available(struct kvm_vcpu *vcpu, gpa_t new_cr3, kvm_vcpu 4387 arch/x86/kvm/mmu.c static bool fast_cr3_switch(struct kvm_vcpu *vcpu, gpa_t new_cr3, kvm_vcpu 4436 arch/x86/kvm/mmu.c static void __kvm_mmu_new_cr3(struct kvm_vcpu *vcpu, gpa_t new_cr3, kvm_vcpu 4445 arch/x86/kvm/mmu.c void kvm_mmu_new_cr3(struct kvm_vcpu *vcpu, gpa_t new_cr3, bool skip_tlb_flush) kvm_vcpu 4452 arch/x86/kvm/mmu.c static unsigned long get_cr3(struct kvm_vcpu *vcpu) kvm_vcpu 4457 arch/x86/kvm/mmu.c static void inject_page_fault(struct kvm_vcpu *vcpu, kvm_vcpu 4463 arch/x86/kvm/mmu.c static bool sync_mmio_spte(struct kvm_vcpu *vcpu, u64 *sptep, gfn_t gfn, kvm_vcpu 4514 arch/x86/kvm/mmu.c __reset_rsvds_bits_mask(struct kvm_vcpu *vcpu, kvm_vcpu 4603 arch/x86/kvm/mmu.c static void reset_rsvds_bits_mask(struct kvm_vcpu *vcpu, kvm_vcpu 4650 arch/x86/kvm/mmu.c static void reset_rsvds_bits_mask_ept(struct kvm_vcpu *vcpu, kvm_vcpu 4663 arch/x86/kvm/mmu.c reset_shadow_zero_bits_mask(struct kvm_vcpu *vcpu, struct kvm_mmu *context) kvm_vcpu 4703 arch/x86/kvm/mmu.c reset_tdp_shadow_zero_bits_mask(struct kvm_vcpu *vcpu, kvm_vcpu 4736 arch/x86/kvm/mmu.c reset_ept_shadow_zero_bits_mask(struct kvm_vcpu *vcpu, kvm_vcpu 4753 arch/x86/kvm/mmu.c static void update_permission_bitmask(struct kvm_vcpu *vcpu, kvm_vcpu 4849 arch/x86/kvm/mmu.c static void update_pkru_bitmask(struct kvm_vcpu *vcpu, struct kvm_mmu *mmu, kvm_vcpu 4900 arch/x86/kvm/mmu.c static void update_last_nonleaf_level(struct kvm_vcpu *vcpu, struct kvm_mmu *mmu) kvm_vcpu 4909 arch/x86/kvm/mmu.c static void paging64_init_context_common(struct kvm_vcpu *vcpu, kvm_vcpu 4931 arch/x86/kvm/mmu.c static void paging64_init_context(struct kvm_vcpu *vcpu, kvm_vcpu 4940 arch/x86/kvm/mmu.c static void paging32_init_context(struct kvm_vcpu *vcpu, kvm_vcpu 4960 arch/x86/kvm/mmu.c static void paging32E_init_context(struct kvm_vcpu *vcpu, kvm_vcpu 4966 arch/x86/kvm/mmu.c static union kvm_mmu_extended_role kvm_calc_mmu_role_ext(struct kvm_vcpu *vcpu) kvm_vcpu 4984 arch/x86/kvm/mmu.c static union kvm_mmu_role kvm_calc_mmu_role_common(struct kvm_vcpu *vcpu, kvm_vcpu 5004 arch/x86/kvm/mmu.c kvm_calc_tdp_mmu_root_page_role(struct kvm_vcpu *vcpu, bool base_only) kvm_vcpu 5016 arch/x86/kvm/mmu.c static void init_kvm_tdp_mmu(struct kvm_vcpu *vcpu) kvm_vcpu 5067 arch/x86/kvm/mmu.c kvm_calc_shadow_mmu_root_page_role(struct kvm_vcpu *vcpu, bool base_only) kvm_vcpu 5088 arch/x86/kvm/mmu.c void kvm_init_shadow_mmu(struct kvm_vcpu *vcpu) kvm_vcpu 5113 arch/x86/kvm/mmu.c kvm_calc_shadow_ept_root_page_role(struct kvm_vcpu *vcpu, bool accessed_dirty, kvm_vcpu 5141 arch/x86/kvm/mmu.c void kvm_init_shadow_ept_mmu(struct kvm_vcpu *vcpu, bool execonly, kvm_vcpu 5176 arch/x86/kvm/mmu.c static void init_kvm_softmmu(struct kvm_vcpu *vcpu) kvm_vcpu 5187 arch/x86/kvm/mmu.c static void init_kvm_nested_mmu(struct kvm_vcpu *vcpu) kvm_vcpu 5236 arch/x86/kvm/mmu.c void kvm_init_mmu(struct kvm_vcpu *vcpu, bool reset_roots) kvm_vcpu 5257 arch/x86/kvm/mmu.c kvm_mmu_calc_root_page_role(struct kvm_vcpu *vcpu) kvm_vcpu 5269 arch/x86/kvm/mmu.c void kvm_mmu_reset_context(struct kvm_vcpu *vcpu) kvm_vcpu 5276 arch/x86/kvm/mmu.c int kvm_mmu_load(struct kvm_vcpu *vcpu) kvm_vcpu 5294 arch/x86/kvm/mmu.c void kvm_mmu_unload(struct kvm_vcpu *vcpu) kvm_vcpu 5303 arch/x86/kvm/mmu.c static void mmu_pte_write_new_pte(struct kvm_vcpu *vcpu, kvm_vcpu 5329 arch/x86/kvm/mmu.c static u64 mmu_pte_write_fetch_gpte(struct kvm_vcpu *vcpu, gpa_t *gpa, kvm_vcpu 5431 arch/x86/kvm/mmu.c static void kvm_mmu_pte_write(struct kvm_vcpu *vcpu, gpa_t gpa, kvm_vcpu 5499 arch/x86/kvm/mmu.c int kvm_mmu_unprotect_page_virt(struct kvm_vcpu *vcpu, gva_t gva) kvm_vcpu 5515 arch/x86/kvm/mmu.c static int make_mmu_pages_available(struct kvm_vcpu *vcpu) kvm_vcpu 5535 arch/x86/kvm/mmu.c int kvm_mmu_page_fault(struct kvm_vcpu *vcpu, gpa_t cr2_or_gpa, u64 error_code, kvm_vcpu 5610 arch/x86/kvm/mmu.c void kvm_mmu_invlpg(struct kvm_vcpu *vcpu, gva_t gva) kvm_vcpu 5641 arch/x86/kvm/mmu.c void kvm_mmu_invpcid_gva(struct kvm_vcpu *vcpu, gva_t gva, unsigned long pcid) kvm_vcpu 5764 arch/x86/kvm/mmu.c static int alloc_mmu_pages(struct kvm_vcpu *vcpu, struct kvm_mmu *mmu) kvm_vcpu 5792 arch/x86/kvm/mmu.c int kvm_mmu_create(struct kvm_vcpu *vcpu) kvm_vcpu 6382 arch/x86/kvm/mmu.c void kvm_mmu_destroy(struct kvm_vcpu *vcpu) kvm_vcpu 57 arch/x86/kvm/mmu.h reset_shadow_zero_bits_mask(struct kvm_vcpu *vcpu, struct kvm_mmu *context); kvm_vcpu 59 arch/x86/kvm/mmu.h void kvm_init_mmu(struct kvm_vcpu *vcpu, bool reset_roots); kvm_vcpu 60 arch/x86/kvm/mmu.h void kvm_init_shadow_mmu(struct kvm_vcpu *vcpu); kvm_vcpu 61 arch/x86/kvm/mmu.h void kvm_init_shadow_ept_mmu(struct kvm_vcpu *vcpu, bool execonly, kvm_vcpu 63 arch/x86/kvm/mmu.h bool kvm_can_do_async_pf(struct kvm_vcpu *vcpu); kvm_vcpu 64 arch/x86/kvm/mmu.h int kvm_handle_page_fault(struct kvm_vcpu *vcpu, u64 error_code, kvm_vcpu 76 arch/x86/kvm/mmu.h static inline int kvm_mmu_reload(struct kvm_vcpu *vcpu) kvm_vcpu 84 arch/x86/kvm/mmu.h static inline unsigned long kvm_get_pcid(struct kvm_vcpu *vcpu, gpa_t cr3) kvm_vcpu 93 arch/x86/kvm/mmu.h static inline unsigned long kvm_get_active_pcid(struct kvm_vcpu *vcpu) kvm_vcpu 98 arch/x86/kvm/mmu.h static inline void kvm_mmu_load_cr3(struct kvm_vcpu *vcpu) kvm_vcpu 143 arch/x86/kvm/mmu.h static inline bool is_write_protection(struct kvm_vcpu *vcpu) kvm_vcpu 156 arch/x86/kvm/mmu.h static inline u8 permission_fault(struct kvm_vcpu *vcpu, struct kvm_mmu *mmu, kvm_vcpu 212 arch/x86/kvm/mmu.h int kvm_arch_write_log_dirty(struct kvm_vcpu *vcpu); kvm_vcpu 32 arch/x86/kvm/mmu_audit.c typedef void (*inspect_spte_fn) (struct kvm_vcpu *vcpu, u64 *sptep, int level); kvm_vcpu 34 arch/x86/kvm/mmu_audit.c static void __mmu_spte_walk(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp, kvm_vcpu 54 arch/x86/kvm/mmu_audit.c static void mmu_spte_walk(struct kvm_vcpu *vcpu, inspect_spte_fn fn) kvm_vcpu 93 arch/x86/kvm/mmu_audit.c static void audit_mappings(struct kvm_vcpu *vcpu, u64 *sptep, int level) kvm_vcpu 160 arch/x86/kvm/mmu_audit.c static void audit_sptes_have_rmaps(struct kvm_vcpu *vcpu, u64 *sptep, int level) kvm_vcpu 166 arch/x86/kvm/mmu_audit.c static void audit_spte_after_sync(struct kvm_vcpu *vcpu, u64 *sptep, int level) kvm_vcpu 224 arch/x86/kvm/mmu_audit.c static void audit_spte(struct kvm_vcpu *vcpu, u64 *sptep, int level) kvm_vcpu 231 arch/x86/kvm/mmu_audit.c static void audit_vcpu_spte(struct kvm_vcpu *vcpu) kvm_vcpu 239 arch/x86/kvm/mmu_audit.c static void __kvm_mmu_audit(struct kvm_vcpu *vcpu, int point) kvm_vcpu 251 arch/x86/kvm/mmu_audit.c static inline void kvm_mmu_audit(struct kvm_vcpu *vcpu, int point) kvm_vcpu 252 arch/x86/kvm/mmutrace.h TP_PROTO(struct kvm_vcpu *vcpu, gpa_t cr2_or_gpa, u32 error_code, kvm_vcpu 54 arch/x86/kvm/mtrr.c bool kvm_mtrr_valid(struct kvm_vcpu *vcpu, u32 msr, u64 data) kvm_vcpu 111 arch/x86/kvm/mtrr.c static u8 mtrr_disabled_type(struct kvm_vcpu *vcpu) kvm_vcpu 310 arch/x86/kvm/mtrr.c static void update_mtrr(struct kvm_vcpu *vcpu, u32 msr) kvm_vcpu 344 arch/x86/kvm/mtrr.c static void set_var_mtrr_msr(struct kvm_vcpu *vcpu, u32 msr, u64 data) kvm_vcpu 376 arch/x86/kvm/mtrr.c int kvm_mtrr_set_msr(struct kvm_vcpu *vcpu, u32 msr, u64 data) kvm_vcpu 397 arch/x86/kvm/mtrr.c int kvm_mtrr_get_msr(struct kvm_vcpu *vcpu, u32 msr, u64 *pdata) kvm_vcpu 439 arch/x86/kvm/mtrr.c void kvm_vcpu_mtrr_init(struct kvm_vcpu *vcpu) kvm_vcpu 619 arch/x86/kvm/mtrr.c u8 kvm_mtrr_get_guest_memory_type(struct kvm_vcpu *vcpu, gfn_t gfn) kvm_vcpu 695 arch/x86/kvm/mtrr.c bool kvm_mtrr_check_gfn_range_consistency(struct kvm_vcpu *vcpu, gfn_t gfn, kvm_vcpu 145 arch/x86/kvm/page_track.c bool kvm_page_track_is_active(struct kvm_vcpu *vcpu, gfn_t gfn, kvm_vcpu 223 arch/x86/kvm/page_track.c void kvm_page_track_write(struct kvm_vcpu *vcpu, gpa_t gpa, const u8 *new, kvm_vcpu 131 arch/x86/kvm/paging_tmpl.h static int FNAME(cmpxchg_gpte)(struct kvm_vcpu *vcpu, struct kvm_mmu *mmu, kvm_vcpu 174 arch/x86/kvm/paging_tmpl.h static bool FNAME(prefetch_invalid_gpte)(struct kvm_vcpu *vcpu, kvm_vcpu 220 arch/x86/kvm/paging_tmpl.h static int FNAME(update_accessed_dirty_bits)(struct kvm_vcpu *vcpu, kvm_vcpu 282 arch/x86/kvm/paging_tmpl.h static inline unsigned FNAME(gpte_pkeys)(struct kvm_vcpu *vcpu, u64 gpte) kvm_vcpu 297 arch/x86/kvm/paging_tmpl.h struct kvm_vcpu *vcpu, struct kvm_mmu *mmu, kvm_vcpu 499 arch/x86/kvm/paging_tmpl.h struct kvm_vcpu *vcpu, gpa_t addr, u32 access) kvm_vcpu 507 arch/x86/kvm/paging_tmpl.h struct kvm_vcpu *vcpu, gva_t addr, kvm_vcpu 516 arch/x86/kvm/paging_tmpl.h FNAME(prefetch_gpte)(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp, kvm_vcpu 547 arch/x86/kvm/paging_tmpl.h static void FNAME(update_pte)(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp, kvm_vcpu 555 arch/x86/kvm/paging_tmpl.h static bool FNAME(gpte_changed)(struct kvm_vcpu *vcpu, kvm_vcpu 578 arch/x86/kvm/paging_tmpl.h static void FNAME(pte_prefetch)(struct kvm_vcpu *vcpu, struct guest_walker *gw, kvm_vcpu 614 arch/x86/kvm/paging_tmpl.h static int FNAME(fetch)(struct kvm_vcpu *vcpu, gpa_t addr, kvm_vcpu 732 arch/x86/kvm/paging_tmpl.h FNAME(is_self_change_mapping)(struct kvm_vcpu *vcpu, kvm_vcpu 768 arch/x86/kvm/paging_tmpl.h static int FNAME(page_fault)(struct kvm_vcpu *vcpu, gpa_t addr, u32 error_code, kvm_vcpu 892 arch/x86/kvm/paging_tmpl.h static void FNAME(invlpg)(struct kvm_vcpu *vcpu, gva_t gva, hpa_t root_hpa) kvm_vcpu 949 arch/x86/kvm/paging_tmpl.h static gpa_t FNAME(gva_to_gpa)(struct kvm_vcpu *vcpu, gpa_t addr, u32 access, kvm_vcpu 969 arch/x86/kvm/paging_tmpl.h static gpa_t FNAME(gva_to_gpa_nested)(struct kvm_vcpu *vcpu, gpa_t vaddr, kvm_vcpu 1007 arch/x86/kvm/paging_tmpl.h static int FNAME(sync_page)(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp) kvm_vcpu 53 arch/x86/kvm/pmu.c struct kvm_vcpu *vcpu = pmu_to_vcpu(pmu); kvm_vcpu 253 arch/x86/kvm/pmu.c void kvm_pmu_handle_event(struct kvm_vcpu *vcpu) kvm_vcpu 274 arch/x86/kvm/pmu.c int kvm_pmu_is_valid_msr_idx(struct kvm_vcpu *vcpu, unsigned idx) kvm_vcpu 290 arch/x86/kvm/pmu.c static int kvm_pmu_rdpmc_vmware(struct kvm_vcpu *vcpu, unsigned idx, u64 *data) kvm_vcpu 313 arch/x86/kvm/pmu.c int kvm_pmu_rdpmc(struct kvm_vcpu *vcpu, unsigned idx, u64 *data) kvm_vcpu 334 arch/x86/kvm/pmu.c void kvm_pmu_deliver_pmi(struct kvm_vcpu *vcpu) kvm_vcpu 340 arch/x86/kvm/pmu.c bool kvm_pmu_is_valid_msr(struct kvm_vcpu *vcpu, u32 msr) kvm_vcpu 345 arch/x86/kvm/pmu.c int kvm_pmu_get_msr(struct kvm_vcpu *vcpu, u32 msr, u64 *data) kvm_vcpu 350 arch/x86/kvm/pmu.c int kvm_pmu_set_msr(struct kvm_vcpu *vcpu, struct msr_data *msr_info) kvm_vcpu 359 arch/x86/kvm/pmu.c void kvm_pmu_refresh(struct kvm_vcpu *vcpu) kvm_vcpu 364 arch/x86/kvm/pmu.c void kvm_pmu_reset(struct kvm_vcpu *vcpu) kvm_vcpu 372 arch/x86/kvm/pmu.c void kvm_pmu_init(struct kvm_vcpu *vcpu) kvm_vcpu 382 arch/x86/kvm/pmu.c void kvm_pmu_destroy(struct kvm_vcpu *vcpu) kvm_vcpu 8 arch/x86/kvm/pmu.h #define pmu_to_vcpu(pmu) (container_of((pmu), struct kvm_vcpu, arch.pmu)) kvm_vcpu 30 arch/x86/kvm/pmu.h struct kvm_pmc *(*msr_idx_to_pmc)(struct kvm_vcpu *vcpu, unsigned idx, kvm_vcpu 32 arch/x86/kvm/pmu.h int (*is_valid_msr_idx)(struct kvm_vcpu *vcpu, unsigned idx); kvm_vcpu 33 arch/x86/kvm/pmu.h bool (*is_valid_msr)(struct kvm_vcpu *vcpu, u32 msr); kvm_vcpu 34 arch/x86/kvm/pmu.h int (*get_msr)(struct kvm_vcpu *vcpu, u32 msr, u64 *data); kvm_vcpu 35 arch/x86/kvm/pmu.h int (*set_msr)(struct kvm_vcpu *vcpu, struct msr_data *msr_info); kvm_vcpu 36 arch/x86/kvm/pmu.h void (*refresh)(struct kvm_vcpu *vcpu); kvm_vcpu 37 arch/x86/kvm/pmu.h void (*init)(struct kvm_vcpu *vcpu); kvm_vcpu 38 arch/x86/kvm/pmu.h void (*reset)(struct kvm_vcpu *vcpu); kvm_vcpu 120 arch/x86/kvm/pmu.h void kvm_pmu_deliver_pmi(struct kvm_vcpu *vcpu); kvm_vcpu 121 arch/x86/kvm/pmu.h void kvm_pmu_handle_event(struct kvm_vcpu *vcpu); kvm_vcpu 122 arch/x86/kvm/pmu.h int kvm_pmu_rdpmc(struct kvm_vcpu *vcpu, unsigned pmc, u64 *data); kvm_vcpu 123 arch/x86/kvm/pmu.h int kvm_pmu_is_valid_msr_idx(struct kvm_vcpu *vcpu, unsigned idx); kvm_vcpu 124 arch/x86/kvm/pmu.h bool kvm_pmu_is_valid_msr(struct kvm_vcpu *vcpu, u32 msr); kvm_vcpu 125 arch/x86/kvm/pmu.h int kvm_pmu_get_msr(struct kvm_vcpu *vcpu, u32 msr, u64 *data); kvm_vcpu 126 arch/x86/kvm/pmu.h int kvm_pmu_set_msr(struct kvm_vcpu *vcpu, struct msr_data *msr_info); kvm_vcpu 127 arch/x86/kvm/pmu.h void kvm_pmu_refresh(struct kvm_vcpu *vcpu); kvm_vcpu 128 arch/x86/kvm/pmu.h void kvm_pmu_reset(struct kvm_vcpu *vcpu); kvm_vcpu 129 arch/x86/kvm/pmu.h void kvm_pmu_init(struct kvm_vcpu *vcpu); kvm_vcpu 130 arch/x86/kvm/pmu.h void kvm_pmu_destroy(struct kvm_vcpu *vcpu); kvm_vcpu 49 arch/x86/kvm/pmu_amd.c struct kvm_vcpu *vcpu = pmu_to_vcpu(pmu); kvm_vcpu 163 arch/x86/kvm/pmu_amd.c struct kvm_vcpu *vcpu = pmu_to_vcpu(pmu); kvm_vcpu 177 arch/x86/kvm/pmu_amd.c static int amd_is_valid_msr_idx(struct kvm_vcpu *vcpu, unsigned idx) kvm_vcpu 187 arch/x86/kvm/pmu_amd.c static struct kvm_pmc *amd_msr_idx_to_pmc(struct kvm_vcpu *vcpu, unsigned idx, u64 *mask) kvm_vcpu 200 arch/x86/kvm/pmu_amd.c static bool amd_is_valid_msr(struct kvm_vcpu *vcpu, u32 msr) kvm_vcpu 211 arch/x86/kvm/pmu_amd.c static int amd_pmu_get_msr(struct kvm_vcpu *vcpu, u32 msr, u64 *data) kvm_vcpu 232 arch/x86/kvm/pmu_amd.c static int amd_pmu_set_msr(struct kvm_vcpu *vcpu, struct msr_data *msr_info) kvm_vcpu 259 arch/x86/kvm/pmu_amd.c static void amd_pmu_refresh(struct kvm_vcpu *vcpu) kvm_vcpu 277 arch/x86/kvm/pmu_amd.c static void amd_pmu_init(struct kvm_vcpu *vcpu) kvm_vcpu 291 arch/x86/kvm/pmu_amd.c static void amd_pmu_reset(struct kvm_vcpu *vcpu) kvm_vcpu 150 arch/x86/kvm/svm.c struct kvm_vcpu; kvm_vcpu 188 arch/x86/kvm/svm.c struct kvm_vcpu vcpu; kvm_vcpu 386 arch/x86/kvm/svm.c static void svm_set_cr0(struct kvm_vcpu *vcpu, unsigned long cr0); kvm_vcpu 387 arch/x86/kvm/svm.c static void svm_flush_tlb(struct kvm_vcpu *vcpu, bool invalidate_gpa); kvm_vcpu 479 arch/x86/kvm/svm.c static inline struct vcpu_svm *to_svm(struct kvm_vcpu *vcpu) kvm_vcpu 490 arch/x86/kvm/svm.c static inline bool avic_vcpu_is_running(struct kvm_vcpu *vcpu) kvm_vcpu 725 arch/x86/kvm/svm.c static int get_npt_level(struct kvm_vcpu *vcpu) kvm_vcpu 734 arch/x86/kvm/svm.c static void svm_set_efer(struct kvm_vcpu *vcpu, u64 efer) kvm_vcpu 756 arch/x86/kvm/svm.c static u32 svm_get_interrupt_shadow(struct kvm_vcpu *vcpu) kvm_vcpu 766 arch/x86/kvm/svm.c static void svm_set_interrupt_shadow(struct kvm_vcpu *vcpu, int mask) kvm_vcpu 777 arch/x86/kvm/svm.c static int skip_emulated_instruction(struct kvm_vcpu *vcpu) kvm_vcpu 800 arch/x86/kvm/svm.c static void svm_queue_exception(struct kvm_vcpu *vcpu) kvm_vcpu 865 arch/x86/kvm/svm.c static void svm_init_osvw(struct kvm_vcpu *vcpu) kvm_vcpu 1041 arch/x86/kvm/svm.c static bool msr_write_intercepted(struct kvm_vcpu *vcpu, unsigned msr) kvm_vcpu 1197 arch/x86/kvm/svm.c struct kvm_vcpu *vcpu = NULL; kvm_vcpu 1264 arch/x86/kvm/svm.c static void grow_ple_window(struct kvm_vcpu *vcpu) kvm_vcpu 1282 arch/x86/kvm/svm.c static void shrink_ple_window(struct kvm_vcpu *vcpu) kvm_vcpu 1490 arch/x86/kvm/svm.c static u64 svm_read_l1_tsc_offset(struct kvm_vcpu *vcpu) kvm_vcpu 1500 arch/x86/kvm/svm.c static u64 svm_write_l1_tsc_offset(struct kvm_vcpu *vcpu, u64 offset) kvm_vcpu 1696 arch/x86/kvm/svm.c static u64 *avic_get_physical_id_entry(struct kvm_vcpu *vcpu, kvm_vcpu 1718 arch/x86/kvm/svm.c static int avic_init_access_page(struct kvm_vcpu *vcpu) kvm_vcpu 1740 arch/x86/kvm/svm.c static int avic_init_backing_page(struct kvm_vcpu *vcpu) kvm_vcpu 2047 arch/x86/kvm/svm.c avic_update_iommu_vcpu_affinity(struct kvm_vcpu *vcpu, int cpu, bool r) kvm_vcpu 2076 arch/x86/kvm/svm.c static void avic_vcpu_load(struct kvm_vcpu *vcpu, int cpu) kvm_vcpu 2108 arch/x86/kvm/svm.c static void avic_vcpu_put(struct kvm_vcpu *vcpu) kvm_vcpu 2127 arch/x86/kvm/svm.c static void avic_set_running(struct kvm_vcpu *vcpu, bool is_run) kvm_vcpu 2138 arch/x86/kvm/svm.c static void svm_vcpu_reset(struct kvm_vcpu *vcpu, bool init_event) kvm_vcpu 2181 arch/x86/kvm/svm.c static struct kvm_vcpu *svm_create_vcpu(struct kvm *kvm, unsigned int id) kvm_vcpu 2291 arch/x86/kvm/svm.c static void svm_free_vcpu(struct kvm_vcpu *vcpu) kvm_vcpu 2312 arch/x86/kvm/svm.c static void svm_vcpu_load(struct kvm_vcpu *vcpu, int cpu) kvm_vcpu 2351 arch/x86/kvm/svm.c static void svm_vcpu_put(struct kvm_vcpu *vcpu) kvm_vcpu 2373 arch/x86/kvm/svm.c static void svm_vcpu_blocking(struct kvm_vcpu *vcpu) kvm_vcpu 2378 arch/x86/kvm/svm.c static void svm_vcpu_unblocking(struct kvm_vcpu *vcpu) kvm_vcpu 2383 arch/x86/kvm/svm.c static unsigned long svm_get_rflags(struct kvm_vcpu *vcpu) kvm_vcpu 2398 arch/x86/kvm/svm.c static void svm_set_rflags(struct kvm_vcpu *vcpu, unsigned long rflags) kvm_vcpu 2411 arch/x86/kvm/svm.c static void svm_cache_reg(struct kvm_vcpu *vcpu, enum kvm_reg reg) kvm_vcpu 2433 arch/x86/kvm/svm.c static struct vmcb_seg *svm_seg(struct kvm_vcpu *vcpu, int seg) kvm_vcpu 2451 arch/x86/kvm/svm.c static u64 svm_get_segment_base(struct kvm_vcpu *vcpu, int seg) kvm_vcpu 2458 arch/x86/kvm/svm.c static void svm_get_segment(struct kvm_vcpu *vcpu, kvm_vcpu 2527 arch/x86/kvm/svm.c static int svm_get_cpl(struct kvm_vcpu *vcpu) kvm_vcpu 2534 arch/x86/kvm/svm.c static void svm_get_idt(struct kvm_vcpu *vcpu, struct desc_ptr *dt) kvm_vcpu 2542 arch/x86/kvm/svm.c static void svm_set_idt(struct kvm_vcpu *vcpu, struct desc_ptr *dt) kvm_vcpu 2551 arch/x86/kvm/svm.c static void svm_get_gdt(struct kvm_vcpu *vcpu, struct desc_ptr *dt) kvm_vcpu 2559 arch/x86/kvm/svm.c static void svm_set_gdt(struct kvm_vcpu *vcpu, struct desc_ptr *dt) kvm_vcpu 2568 arch/x86/kvm/svm.c static void svm_decache_cr0_guest_bits(struct kvm_vcpu *vcpu) kvm_vcpu 2572 arch/x86/kvm/svm.c static void svm_decache_cr3(struct kvm_vcpu *vcpu) kvm_vcpu 2576 arch/x86/kvm/svm.c static void svm_decache_cr4_guest_bits(struct kvm_vcpu *vcpu) kvm_vcpu 2599 arch/x86/kvm/svm.c static void svm_set_cr0(struct kvm_vcpu *vcpu, unsigned long cr0) kvm_vcpu 2633 arch/x86/kvm/svm.c static int svm_set_cr4(struct kvm_vcpu *vcpu, unsigned long cr4) kvm_vcpu 2653 arch/x86/kvm/svm.c static void svm_set_segment(struct kvm_vcpu *vcpu, kvm_vcpu 2684 arch/x86/kvm/svm.c static void update_bp_intercept(struct kvm_vcpu *vcpu) kvm_vcpu 2711 arch/x86/kvm/svm.c static u64 svm_get_dr6(struct kvm_vcpu *vcpu) kvm_vcpu 2716 arch/x86/kvm/svm.c static void svm_set_dr6(struct kvm_vcpu *vcpu, unsigned long value) kvm_vcpu 2724 arch/x86/kvm/svm.c static void svm_sync_dirty_debug_regs(struct kvm_vcpu *vcpu) kvm_vcpu 2739 arch/x86/kvm/svm.c static void svm_set_dr7(struct kvm_vcpu *vcpu, unsigned long value) kvm_vcpu 2773 arch/x86/kvm/svm.c struct kvm_vcpu *vcpu = &svm->vcpu; kvm_vcpu 2823 arch/x86/kvm/svm.c struct kvm_vcpu *vcpu = &svm->vcpu; kvm_vcpu 2925 arch/x86/kvm/svm.c struct kvm_vcpu *vcpu = &svm->vcpu; kvm_vcpu 2969 arch/x86/kvm/svm.c static unsigned long nested_svm_get_tdp_cr3(struct kvm_vcpu *vcpu) kvm_vcpu 2976 arch/x86/kvm/svm.c static u64 nested_svm_get_tdp_pdptr(struct kvm_vcpu *vcpu, int index) kvm_vcpu 2990 arch/x86/kvm/svm.c static void nested_svm_set_tdp_cr3(struct kvm_vcpu *vcpu, kvm_vcpu 2999 arch/x86/kvm/svm.c static void nested_svm_inject_npf_exit(struct kvm_vcpu *vcpu, kvm_vcpu 3028 arch/x86/kvm/svm.c static void nested_svm_init_mmu_context(struct kvm_vcpu *vcpu) kvm_vcpu 3043 arch/x86/kvm/svm.c static void nested_svm_uninit_mmu_context(struct kvm_vcpu *vcpu) kvm_vcpu 3852 arch/x86/kvm/svm.c struct kvm_vcpu *vcpu = &svm->vcpu; kvm_vcpu 4169 arch/x86/kvm/svm.c static int svm_get_msr(struct kvm_vcpu *vcpu, struct msr_data *msr_info) kvm_vcpu 4277 arch/x86/kvm/svm.c static int svm_set_vm_cr(struct kvm_vcpu *vcpu, u64 data) kvm_vcpu 4302 arch/x86/kvm/svm.c static int svm_set_msr(struct kvm_vcpu *vcpu, struct msr_data *msr) kvm_vcpu 4487 arch/x86/kvm/svm.c struct kvm_vcpu *vcpu = &svm->vcpu; kvm_vcpu 4549 arch/x86/kvm/svm.c struct kvm_vcpu *vcpu; kvm_vcpu 4583 arch/x86/kvm/svm.c static u32 *avic_get_logical_id_entry(struct kvm_vcpu *vcpu, u32 ldr, bool flat) kvm_vcpu 4612 arch/x86/kvm/svm.c static int avic_ldr_write(struct kvm_vcpu *vcpu, u8 g_physical_id, u32 ldr) kvm_vcpu 4631 arch/x86/kvm/svm.c static void avic_invalidate_logical_id_entry(struct kvm_vcpu *vcpu) kvm_vcpu 4641 arch/x86/kvm/svm.c static int avic_handle_ldr_update(struct kvm_vcpu *vcpu) kvm_vcpu 4662 arch/x86/kvm/svm.c static int avic_handle_apic_id_update(struct kvm_vcpu *vcpu) kvm_vcpu 4691 arch/x86/kvm/svm.c static void avic_handle_dfr_update(struct kvm_vcpu *vcpu) kvm_vcpu 4852 arch/x86/kvm/svm.c static void dump_vmcb(struct kvm_vcpu *vcpu) kvm_vcpu 4968 arch/x86/kvm/svm.c static void svm_get_exit_info(struct kvm_vcpu *vcpu, u64 *info1, u64 *info2) kvm_vcpu 4976 arch/x86/kvm/svm.c static int handle_exit(struct kvm_vcpu *vcpu) kvm_vcpu 5049 arch/x86/kvm/svm.c static void reload_tss(struct kvm_vcpu *vcpu) kvm_vcpu 5096 arch/x86/kvm/svm.c static void svm_inject_nmi(struct kvm_vcpu *vcpu) kvm_vcpu 5119 arch/x86/kvm/svm.c static void svm_set_irq(struct kvm_vcpu *vcpu) kvm_vcpu 5132 arch/x86/kvm/svm.c static inline bool svm_nested_virtualize_tpr(struct kvm_vcpu *vcpu) kvm_vcpu 5137 arch/x86/kvm/svm.c static void update_cr8_intercept(struct kvm_vcpu *vcpu, int tpr, int irr) kvm_vcpu 5154 arch/x86/kvm/svm.c static void svm_set_virtual_apic_mode(struct kvm_vcpu *vcpu) kvm_vcpu 5159 arch/x86/kvm/svm.c static bool svm_get_enable_apicv(struct kvm_vcpu *vcpu) kvm_vcpu 5164 arch/x86/kvm/svm.c static void svm_hwapic_irr_update(struct kvm_vcpu *vcpu, int max_irr) kvm_vcpu 5168 arch/x86/kvm/svm.c static void svm_hwapic_isr_update(struct kvm_vcpu *vcpu, int max_isr) kvm_vcpu 5173 arch/x86/kvm/svm.c static void svm_refresh_apicv_exec_ctrl(struct kvm_vcpu *vcpu) kvm_vcpu 5185 arch/x86/kvm/svm.c static void svm_load_eoi_exitmap(struct kvm_vcpu *vcpu, u64 *eoi_exit_bitmap) kvm_vcpu 5190 arch/x86/kvm/svm.c static int svm_deliver_avic_intr(struct kvm_vcpu *vcpu, int vec) kvm_vcpu 5210 arch/x86/kvm/svm.c static bool svm_dy_apicv_has_pending_interrupt(struct kvm_vcpu *vcpu) kvm_vcpu 5245 arch/x86/kvm/svm.c struct kvm_vcpu *prev_vcpu = kvm_get_vcpu_by_id(kvm, vcpu_id); kvm_vcpu 5291 arch/x86/kvm/svm.c struct kvm_vcpu *vcpu = NULL; kvm_vcpu 5394 arch/x86/kvm/svm.c struct kvm_vcpu *vcpu; kvm_vcpu 5420 arch/x86/kvm/svm.c static int svm_nmi_allowed(struct kvm_vcpu *vcpu) kvm_vcpu 5432 arch/x86/kvm/svm.c static bool svm_get_nmi_mask(struct kvm_vcpu *vcpu) kvm_vcpu 5439 arch/x86/kvm/svm.c static void svm_set_nmi_mask(struct kvm_vcpu *vcpu, bool masked) kvm_vcpu 5452 arch/x86/kvm/svm.c static int svm_interrupt_allowed(struct kvm_vcpu *vcpu) kvm_vcpu 5470 arch/x86/kvm/svm.c static void enable_irq_window(struct kvm_vcpu *vcpu) kvm_vcpu 5491 arch/x86/kvm/svm.c static void enable_nmi_window(struct kvm_vcpu *vcpu) kvm_vcpu 5527 arch/x86/kvm/svm.c static void svm_flush_tlb(struct kvm_vcpu *vcpu, bool invalidate_gpa) kvm_vcpu 5537 arch/x86/kvm/svm.c static void svm_flush_tlb_gva(struct kvm_vcpu *vcpu, gva_t gva) kvm_vcpu 5544 arch/x86/kvm/svm.c static void svm_prepare_guest_switch(struct kvm_vcpu *vcpu) kvm_vcpu 5548 arch/x86/kvm/svm.c static inline void sync_cr8_to_lapic(struct kvm_vcpu *vcpu) kvm_vcpu 5561 arch/x86/kvm/svm.c static inline void sync_lapic_to_cr8(struct kvm_vcpu *vcpu) kvm_vcpu 5639 arch/x86/kvm/svm.c static void svm_cancel_injection(struct kvm_vcpu *vcpu) kvm_vcpu 5650 arch/x86/kvm/svm.c static void svm_vcpu_run(struct kvm_vcpu *vcpu) kvm_vcpu 5873 arch/x86/kvm/svm.c static void svm_set_cr3(struct kvm_vcpu *vcpu, unsigned long root) kvm_vcpu 5881 arch/x86/kvm/svm.c static void set_tdp_cr3(struct kvm_vcpu *vcpu, unsigned long root) kvm_vcpu 5905 arch/x86/kvm/svm.c svm_patch_hypercall(struct kvm_vcpu *vcpu, unsigned char *hypercall) kvm_vcpu 5938 arch/x86/kvm/svm.c static u64 svm_get_mt_mask(struct kvm_vcpu *vcpu, gfn_t gfn, bool is_mmio) kvm_vcpu 5943 arch/x86/kvm/svm.c static void svm_cpuid_update(struct kvm_vcpu *vcpu) kvm_vcpu 6109 arch/x86/kvm/svm.c static int svm_check_intercept(struct kvm_vcpu *vcpu, kvm_vcpu 6230 arch/x86/kvm/svm.c static void svm_handle_exit_irqoff(struct kvm_vcpu *vcpu) kvm_vcpu 6235 arch/x86/kvm/svm.c static void svm_sched_in(struct kvm_vcpu *vcpu, int cpu) kvm_vcpu 6241 arch/x86/kvm/svm.c static inline void avic_post_state_restore(struct kvm_vcpu *vcpu) kvm_vcpu 6249 arch/x86/kvm/svm.c static void svm_setup_mce(struct kvm_vcpu *vcpu) kvm_vcpu 6255 arch/x86/kvm/svm.c static int svm_smi_allowed(struct kvm_vcpu *vcpu) kvm_vcpu 6274 arch/x86/kvm/svm.c static int svm_pre_enter_smm(struct kvm_vcpu *vcpu, char *smstate) kvm_vcpu 6296 arch/x86/kvm/svm.c static int svm_pre_leave_smm(struct kvm_vcpu *vcpu, const char *smstate) kvm_vcpu 6316 arch/x86/kvm/svm.c static int enable_smi_window(struct kvm_vcpu *vcpu) kvm_vcpu 7164 arch/x86/kvm/svm.c static bool svm_need_emulation_on_page_fault(struct kvm_vcpu *vcpu) kvm_vcpu 7222 arch/x86/kvm/svm.c static bool svm_apic_init_signal_blocked(struct kvm_vcpu *vcpu) kvm_vcpu 226 arch/x86/kvm/trace.h TP_PROTO(unsigned int exit_reason, struct kvm_vcpu *vcpu, u32 isa), kvm_vcpu 734 arch/x86/kvm/trace.h TP_PROTO(struct kvm_vcpu *vcpu, __u8 failed), kvm_vcpu 115 arch/x86/kvm/vmx/capabilities.h static inline bool cpu_need_tpr_shadow(struct kvm_vcpu *vcpu) kvm_vcpu 317 arch/x86/kvm/vmx/evmcs.c bool nested_enlightened_vmentry(struct kvm_vcpu *vcpu, u64 *evmcs_gpa) kvm_vcpu 334 arch/x86/kvm/vmx/evmcs.c uint16_t nested_get_evmcs_version(struct kvm_vcpu *vcpu) kvm_vcpu 349 arch/x86/kvm/vmx/evmcs.c int nested_enable_evmcs(struct kvm_vcpu *vcpu, kvm_vcpu 200 arch/x86/kvm/vmx/evmcs.h bool nested_enlightened_vmentry(struct kvm_vcpu *vcpu, u64 *evmcs_gpa); kvm_vcpu 201 arch/x86/kvm/vmx/evmcs.h uint16_t nested_get_evmcs_version(struct kvm_vcpu *vcpu); kvm_vcpu 202 arch/x86/kvm/vmx/evmcs.h int nested_enable_evmcs(struct kvm_vcpu *vcpu, kvm_vcpu 153 arch/x86/kvm/vmx/nested.c static int nested_vmx_succeed(struct kvm_vcpu *vcpu) kvm_vcpu 161 arch/x86/kvm/vmx/nested.c static int nested_vmx_failInvalid(struct kvm_vcpu *vcpu) kvm_vcpu 170 arch/x86/kvm/vmx/nested.c static int nested_vmx_failValid(struct kvm_vcpu *vcpu, kvm_vcpu 194 arch/x86/kvm/vmx/nested.c static void nested_vmx_abort(struct kvm_vcpu *vcpu, u32 indicator) kvm_vcpu 218 arch/x86/kvm/vmx/nested.c static inline void nested_release_evmcs(struct kvm_vcpu *vcpu) kvm_vcpu 234 arch/x86/kvm/vmx/nested.c static void free_nested(struct kvm_vcpu *vcpu) kvm_vcpu 293 arch/x86/kvm/vmx/nested.c static void vmx_switch_vmcs(struct kvm_vcpu *vcpu, struct loaded_vmcs *vmcs) kvm_vcpu 316 arch/x86/kvm/vmx/nested.c void nested_vmx_free_vcpu(struct kvm_vcpu *vcpu) kvm_vcpu 325 arch/x86/kvm/vmx/nested.c static void nested_ept_inject_page_fault(struct kvm_vcpu *vcpu, kvm_vcpu 346 arch/x86/kvm/vmx/nested.c static void nested_ept_init_mmu_context(struct kvm_vcpu *vcpu) kvm_vcpu 364 arch/x86/kvm/vmx/nested.c static void nested_ept_uninit_mmu_context(struct kvm_vcpu *vcpu) kvm_vcpu 387 arch/x86/kvm/vmx/nested.c static int nested_vmx_check_exception(struct kvm_vcpu *vcpu, unsigned long *exit_qual) kvm_vcpu 421 arch/x86/kvm/vmx/nested.c static void vmx_inject_page_fault_nested(struct kvm_vcpu *vcpu, kvm_vcpu 440 arch/x86/kvm/vmx/nested.c static bool page_address_valid(struct kvm_vcpu *vcpu, gpa_t gpa) kvm_vcpu 445 arch/x86/kvm/vmx/nested.c static int nested_vmx_check_io_bitmap_controls(struct kvm_vcpu *vcpu, kvm_vcpu 458 arch/x86/kvm/vmx/nested.c static int nested_vmx_check_msr_bitmap_controls(struct kvm_vcpu *vcpu, kvm_vcpu 470 arch/x86/kvm/vmx/nested.c static int nested_vmx_check_tpr_shadow_controls(struct kvm_vcpu *vcpu, kvm_vcpu 485 arch/x86/kvm/vmx/nested.c static bool msr_write_intercepted_l01(struct kvm_vcpu *vcpu, u32 msr) kvm_vcpu 561 arch/x86/kvm/vmx/nested.c static inline bool nested_vmx_prepare_msr_bitmap(struct kvm_vcpu *vcpu, kvm_vcpu 658 arch/x86/kvm/vmx/nested.c static void nested_cache_shadow_vmcs12(struct kvm_vcpu *vcpu, kvm_vcpu 677 arch/x86/kvm/vmx/nested.c static void nested_flush_cached_shadow_vmcs12(struct kvm_vcpu *vcpu, kvm_vcpu 694 arch/x86/kvm/vmx/nested.c static bool nested_exit_intr_ack_set(struct kvm_vcpu *vcpu) kvm_vcpu 700 arch/x86/kvm/vmx/nested.c static bool nested_exit_on_nmi(struct kvm_vcpu *vcpu) kvm_vcpu 705 arch/x86/kvm/vmx/nested.c static int nested_vmx_check_apic_access_controls(struct kvm_vcpu *vcpu, kvm_vcpu 715 arch/x86/kvm/vmx/nested.c static int nested_vmx_check_apicv_controls(struct kvm_vcpu *vcpu, kvm_vcpu 761 arch/x86/kvm/vmx/nested.c static int nested_vmx_check_msr_switch(struct kvm_vcpu *vcpu, kvm_vcpu 776 arch/x86/kvm/vmx/nested.c static int nested_vmx_check_exit_msr_switch_controls(struct kvm_vcpu *vcpu, kvm_vcpu 790 arch/x86/kvm/vmx/nested.c static int nested_vmx_check_entry_msr_switch_controls(struct kvm_vcpu *vcpu, kvm_vcpu 801 arch/x86/kvm/vmx/nested.c static int nested_vmx_check_pml_controls(struct kvm_vcpu *vcpu, kvm_vcpu 814 arch/x86/kvm/vmx/nested.c static int nested_vmx_check_unrestricted_guest_controls(struct kvm_vcpu *vcpu, kvm_vcpu 823 arch/x86/kvm/vmx/nested.c static int nested_vmx_check_mode_based_ept_exec_controls(struct kvm_vcpu *vcpu, kvm_vcpu 832 arch/x86/kvm/vmx/nested.c static int nested_vmx_check_shadow_vmcs_controls(struct kvm_vcpu *vcpu, kvm_vcpu 845 arch/x86/kvm/vmx/nested.c static int nested_vmx_msr_check_common(struct kvm_vcpu *vcpu, kvm_vcpu 859 arch/x86/kvm/vmx/nested.c static int nested_vmx_load_msr_check(struct kvm_vcpu *vcpu, kvm_vcpu 870 arch/x86/kvm/vmx/nested.c static int nested_vmx_store_msr_check(struct kvm_vcpu *vcpu, kvm_vcpu 879 arch/x86/kvm/vmx/nested.c static u32 nested_vmx_max_atomic_switch_msrs(struct kvm_vcpu *vcpu) kvm_vcpu 897 arch/x86/kvm/vmx/nested.c static u32 nested_vmx_load_msr(struct kvm_vcpu *vcpu, u64 gpa, u32 count) kvm_vcpu 932 arch/x86/kvm/vmx/nested.c static int nested_vmx_store_msr(struct kvm_vcpu *vcpu, u64 gpa, u32 count) kvm_vcpu 976 arch/x86/kvm/vmx/nested.c static bool nested_cr3_valid(struct kvm_vcpu *vcpu, unsigned long val) kvm_vcpu 990 arch/x86/kvm/vmx/nested.c static int nested_vmx_load_cr3(struct kvm_vcpu *vcpu, unsigned long cr3, bool nested_ept, kvm_vcpu 1033 arch/x86/kvm/vmx/nested.c static bool nested_has_guest_tlb_tag(struct kvm_vcpu *vcpu) kvm_vcpu 1041 arch/x86/kvm/vmx/nested.c static u16 nested_get_vpid02(struct kvm_vcpu *vcpu) kvm_vcpu 1216 arch/x86/kvm/vmx/nested.c int vmx_set_vmx_msr(struct kvm_vcpu *vcpu, u32 msr_index, u64 data) kvm_vcpu 1818 arch/x86/kvm/vmx/nested.c static int nested_vmx_handle_enlightened_vmptrld(struct kvm_vcpu *vcpu, kvm_vcpu 1901 arch/x86/kvm/vmx/nested.c void nested_sync_vmcs12_to_shadow(struct kvm_vcpu *vcpu) kvm_vcpu 1937 arch/x86/kvm/vmx/nested.c static void vmx_start_preemption_timer(struct kvm_vcpu *vcpu) kvm_vcpu 2306 arch/x86/kvm/vmx/nested.c static int prepare_vmcs02(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12, kvm_vcpu 2462 arch/x86/kvm/vmx/nested.c static bool valid_ept_address(struct kvm_vcpu *vcpu, u64 address) kvm_vcpu 2501 arch/x86/kvm/vmx/nested.c static int nested_check_vm_execution_controls(struct kvm_vcpu *vcpu, kvm_vcpu 2560 arch/x86/kvm/vmx/nested.c static int nested_check_vm_exit_controls(struct kvm_vcpu *vcpu, kvm_vcpu 2577 arch/x86/kvm/vmx/nested.c static int nested_check_vm_entry_controls(struct kvm_vcpu *vcpu, kvm_vcpu 2649 arch/x86/kvm/vmx/nested.c static int nested_vmx_check_controls(struct kvm_vcpu *vcpu, kvm_vcpu 2660 arch/x86/kvm/vmx/nested.c static int nested_vmx_check_host_state(struct kvm_vcpu *vcpu, kvm_vcpu 2734 arch/x86/kvm/vmx/nested.c static int nested_vmx_check_vmcs_link_ptr(struct kvm_vcpu *vcpu, kvm_vcpu 2772 arch/x86/kvm/vmx/nested.c static int nested_vmx_check_guest_state(struct kvm_vcpu *vcpu, kvm_vcpu 2823 arch/x86/kvm/vmx/nested.c static int nested_vmx_check_vmentry_hw(struct kvm_vcpu *vcpu) kvm_vcpu 2928 arch/x86/kvm/vmx/nested.c static inline bool nested_vmx_prepare_msr_bitmap(struct kvm_vcpu *vcpu, kvm_vcpu 2931 arch/x86/kvm/vmx/nested.c static bool nested_get_vmcs12_pages(struct kvm_vcpu *vcpu) kvm_vcpu 3017 arch/x86/kvm/vmx/nested.c static int nested_vmx_check_permission(struct kvm_vcpu *vcpu) kvm_vcpu 3032 arch/x86/kvm/vmx/nested.c static u8 vmx_has_apicv_interrupt(struct kvm_vcpu *vcpu) kvm_vcpu 3040 arch/x86/kvm/vmx/nested.c static void load_vmcs12_host_state(struct kvm_vcpu *vcpu, kvm_vcpu 3053 arch/x86/kvm/vmx/nested.c enum nvmx_vmentry_status nested_vmx_enter_non_root_mode(struct kvm_vcpu *vcpu, kvm_vcpu 3196 arch/x86/kvm/vmx/nested.c static int nested_vmx_run(struct kvm_vcpu *vcpu, bool launch) kvm_vcpu 3323 arch/x86/kvm/vmx/nested.c vmcs12_guest_cr0(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) kvm_vcpu 3333 arch/x86/kvm/vmx/nested.c vmcs12_guest_cr4(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) kvm_vcpu 3342 arch/x86/kvm/vmx/nested.c static void vmcs12_save_pending_event(struct kvm_vcpu *vcpu, kvm_vcpu 3385 arch/x86/kvm/vmx/nested.c static void nested_mark_vmcs12_pages_dirty(struct kvm_vcpu *vcpu) kvm_vcpu 3406 arch/x86/kvm/vmx/nested.c static void vmx_complete_nested_posted_interrupt(struct kvm_vcpu *vcpu) kvm_vcpu 3439 arch/x86/kvm/vmx/nested.c static void nested_vmx_inject_exception_vmexit(struct kvm_vcpu *vcpu, kvm_vcpu 3463 arch/x86/kvm/vmx/nested.c static int vmx_check_nested_events(struct kvm_vcpu *vcpu) kvm_vcpu 3521 arch/x86/kvm/vmx/nested.c static u32 vmx_get_preemption_timer_value(struct kvm_vcpu *vcpu) kvm_vcpu 3582 arch/x86/kvm/vmx/nested.c static void sync_vmcs02_to_vmcs12_rare(struct kvm_vcpu *vcpu, kvm_vcpu 3629 arch/x86/kvm/vmx/nested.c static void copy_vmcs02_to_vmcs12_rare(struct kvm_vcpu *vcpu, kvm_vcpu 3658 arch/x86/kvm/vmx/nested.c static void sync_vmcs02_to_vmcs12(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) kvm_vcpu 3739 arch/x86/kvm/vmx/nested.c static void prepare_vmcs12(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12, kvm_vcpu 3796 arch/x86/kvm/vmx/nested.c static void load_vmcs12_host_state(struct kvm_vcpu *vcpu, kvm_vcpu 3960 arch/x86/kvm/vmx/nested.c static void nested_vmx_restore_host_state(struct kvm_vcpu *vcpu) kvm_vcpu 4073 arch/x86/kvm/vmx/nested.c void nested_vmx_vmexit(struct kvm_vcpu *vcpu, u32 exit_reason, kvm_vcpu 4207 arch/x86/kvm/vmx/nested.c int get_vmx_mem_address(struct kvm_vcpu *vcpu, unsigned long exit_qualification, kvm_vcpu 4331 arch/x86/kvm/vmx/nested.c static int nested_vmx_get_vmptr(struct kvm_vcpu *vcpu, gpa_t *vmpointer) kvm_vcpu 4354 arch/x86/kvm/vmx/nested.c static struct vmcs *alloc_shadow_vmcs(struct kvm_vcpu *vcpu) kvm_vcpu 4375 arch/x86/kvm/vmx/nested.c static int enter_vmx_operation(struct kvm_vcpu *vcpu) kvm_vcpu 4432 arch/x86/kvm/vmx/nested.c static int handle_vmon(struct kvm_vcpu *vcpu) kvm_vcpu 4497 arch/x86/kvm/vmx/nested.c static inline void nested_release_vmcs12(struct kvm_vcpu *vcpu) kvm_vcpu 4525 arch/x86/kvm/vmx/nested.c static int handle_vmoff(struct kvm_vcpu *vcpu) kvm_vcpu 4539 arch/x86/kvm/vmx/nested.c static int handle_vmclear(struct kvm_vcpu *vcpu) kvm_vcpu 4584 arch/x86/kvm/vmx/nested.c static int nested_vmx_run(struct kvm_vcpu *vcpu, bool launch); kvm_vcpu 4587 arch/x86/kvm/vmx/nested.c static int handle_vmlaunch(struct kvm_vcpu *vcpu) kvm_vcpu 4593 arch/x86/kvm/vmx/nested.c static int handle_vmresume(struct kvm_vcpu *vcpu) kvm_vcpu 4599 arch/x86/kvm/vmx/nested.c static int handle_vmread(struct kvm_vcpu *vcpu) kvm_vcpu 4686 arch/x86/kvm/vmx/nested.c static int handle_vmwrite(struct kvm_vcpu *vcpu) kvm_vcpu 4810 arch/x86/kvm/vmx/nested.c static int handle_vmptrld(struct kvm_vcpu *vcpu) kvm_vcpu 4874 arch/x86/kvm/vmx/nested.c static int handle_vmptrst(struct kvm_vcpu *vcpu) kvm_vcpu 4901 arch/x86/kvm/vmx/nested.c static int handle_invept(struct kvm_vcpu *vcpu) kvm_vcpu 4958 arch/x86/kvm/vmx/nested.c static int handle_invvpid(struct kvm_vcpu *vcpu) kvm_vcpu 5036 arch/x86/kvm/vmx/nested.c static int nested_vmx_eptp_switching(struct kvm_vcpu *vcpu, kvm_vcpu 5081 arch/x86/kvm/vmx/nested.c static int handle_vmfunc(struct kvm_vcpu *vcpu) kvm_vcpu 5122 arch/x86/kvm/vmx/nested.c bool nested_vmx_check_io_bitmaps(struct kvm_vcpu *vcpu, unsigned int port, kvm_vcpu 5155 arch/x86/kvm/vmx/nested.c static bool nested_vmx_exit_handled_io(struct kvm_vcpu *vcpu, kvm_vcpu 5179 arch/x86/kvm/vmx/nested.c static bool nested_vmx_exit_handled_msr(struct kvm_vcpu *vcpu, kvm_vcpu 5216 arch/x86/kvm/vmx/nested.c static bool nested_vmx_exit_handled_cr(struct kvm_vcpu *vcpu, kvm_vcpu 5295 arch/x86/kvm/vmx/nested.c static bool nested_vmx_exit_handled_vmcs_access(struct kvm_vcpu *vcpu, kvm_vcpu 5324 arch/x86/kvm/vmx/nested.c bool nested_vmx_exit_reflected(struct kvm_vcpu *vcpu, u32 exit_reason) kvm_vcpu 5506 arch/x86/kvm/vmx/nested.c static int vmx_get_nested_state(struct kvm_vcpu *vcpu, kvm_vcpu 5609 arch/x86/kvm/vmx/nested.c void vmx_leave_nested(struct kvm_vcpu *vcpu) kvm_vcpu 5618 arch/x86/kvm/vmx/nested.c static int vmx_set_nested_state(struct kvm_vcpu *vcpu, kvm_vcpu 6032 arch/x86/kvm/vmx/nested.c __init int nested_vmx_hardware_setup(int (*exit_handlers[])(struct kvm_vcpu *)) kvm_vcpu 19 arch/x86/kvm/vmx/nested.h void vmx_leave_nested(struct kvm_vcpu *vcpu); kvm_vcpu 22 arch/x86/kvm/vmx/nested.h __init int nested_vmx_hardware_setup(int (*exit_handlers[])(struct kvm_vcpu *)); kvm_vcpu 24 arch/x86/kvm/vmx/nested.h void nested_vmx_free_vcpu(struct kvm_vcpu *vcpu); kvm_vcpu 25 arch/x86/kvm/vmx/nested.h enum nvmx_vmentry_status nested_vmx_enter_non_root_mode(struct kvm_vcpu *vcpu, kvm_vcpu 27 arch/x86/kvm/vmx/nested.h bool nested_vmx_exit_reflected(struct kvm_vcpu *vcpu, u32 exit_reason); kvm_vcpu 28 arch/x86/kvm/vmx/nested.h void nested_vmx_vmexit(struct kvm_vcpu *vcpu, u32 exit_reason, kvm_vcpu 30 arch/x86/kvm/vmx/nested.h void nested_sync_vmcs12_to_shadow(struct kvm_vcpu *vcpu); kvm_vcpu 31 arch/x86/kvm/vmx/nested.h int vmx_set_vmx_msr(struct kvm_vcpu *vcpu, u32 msr_index, u64 data); kvm_vcpu 33 arch/x86/kvm/vmx/nested.h int get_vmx_mem_address(struct kvm_vcpu *vcpu, unsigned long exit_qualification, kvm_vcpu 35 arch/x86/kvm/vmx/nested.h bool nested_vmx_check_io_bitmaps(struct kvm_vcpu *vcpu, unsigned int port, kvm_vcpu 38 arch/x86/kvm/vmx/nested.h static inline struct vmcs12 *get_vmcs12(struct kvm_vcpu *vcpu) kvm_vcpu 43 arch/x86/kvm/vmx/nested.h static inline struct vmcs12 *get_shadow_vmcs12(struct kvm_vcpu *vcpu) kvm_vcpu 48 arch/x86/kvm/vmx/nested.h static inline int vmx_has_valid_vmcs12(struct kvm_vcpu *vcpu) kvm_vcpu 62 arch/x86/kvm/vmx/nested.h static inline unsigned long nested_ept_get_cr3(struct kvm_vcpu *vcpu) kvm_vcpu 68 arch/x86/kvm/vmx/nested.h static inline bool nested_ept_ad_enabled(struct kvm_vcpu *vcpu) kvm_vcpu 76 arch/x86/kvm/vmx/nested.h static inline int nested_vmx_reflect_vmexit(struct kvm_vcpu *vcpu, kvm_vcpu 117 arch/x86/kvm/vmx/nested.h static inline unsigned nested_cpu_vmx_misc_cr3_count(struct kvm_vcpu *vcpu) kvm_vcpu 127 arch/x86/kvm/vmx/nested.h static inline bool nested_cpu_has_vmwrite_any_field(struct kvm_vcpu *vcpu) kvm_vcpu 133 arch/x86/kvm/vmx/nested.h static inline bool nested_cpu_has_zero_length_injection(struct kvm_vcpu *vcpu) kvm_vcpu 138 arch/x86/kvm/vmx/nested.h static inline bool nested_cpu_supports_monitor_trap_flag(struct kvm_vcpu *vcpu) kvm_vcpu 144 arch/x86/kvm/vmx/nested.h static inline bool nested_cpu_has_vmx_shadow_vmcs(struct kvm_vcpu *vcpu) kvm_vcpu 245 arch/x86/kvm/vmx/nested.h static inline bool nested_exit_on_intr(struct kvm_vcpu *vcpu) kvm_vcpu 260 arch/x86/kvm/vmx/nested.h static bool nested_guest_cr0_valid(struct kvm_vcpu *vcpu, unsigned long val) kvm_vcpu 274 arch/x86/kvm/vmx/nested.h static bool nested_host_cr0_valid(struct kvm_vcpu *vcpu, unsigned long val) kvm_vcpu 282 arch/x86/kvm/vmx/nested.h static bool nested_cr4_valid(struct kvm_vcpu *vcpu, unsigned long val) kvm_vcpu 118 arch/x86/kvm/vmx/pmu_intel.c static int intel_is_valid_msr_idx(struct kvm_vcpu *vcpu, unsigned idx) kvm_vcpu 129 arch/x86/kvm/vmx/pmu_intel.c static struct kvm_pmc *intel_msr_idx_to_pmc(struct kvm_vcpu *vcpu, kvm_vcpu 151 arch/x86/kvm/vmx/pmu_intel.c static bool intel_is_valid_msr(struct kvm_vcpu *vcpu, u32 msr) kvm_vcpu 173 arch/x86/kvm/vmx/pmu_intel.c static int intel_pmu_get_msr(struct kvm_vcpu *vcpu, u32 msr, u64 *data) kvm_vcpu 209 arch/x86/kvm/vmx/pmu_intel.c static int intel_pmu_set_msr(struct kvm_vcpu *vcpu, struct msr_data *msr_info) kvm_vcpu 270 arch/x86/kvm/vmx/pmu_intel.c static void intel_pmu_refresh(struct kvm_vcpu *vcpu) kvm_vcpu 330 arch/x86/kvm/vmx/pmu_intel.c static void intel_pmu_init(struct kvm_vcpu *vcpu) kvm_vcpu 348 arch/x86/kvm/vmx/pmu_intel.c static void intel_pmu_reset(struct kvm_vcpu *vcpu) kvm_vcpu 344 arch/x86/kvm/vmx/vmx.c static bool guest_state_valid(struct kvm_vcpu *vcpu); kvm_vcpu 462 arch/x86/kvm/vmx/vmx.c struct kvm_vcpu *vcpu; kvm_vcpu 489 arch/x86/kvm/vmx/vmx.c struct kvm_vcpu *vcpu, struct kvm_tlb_range *range) kvm_vcpu 508 arch/x86/kvm/vmx/vmx.c struct kvm_vcpu *vcpu; kvm_vcpu 536 arch/x86/kvm/vmx/vmx.c static int hv_enable_direct_tlbflush(struct kvm_vcpu *vcpu) kvm_vcpu 611 arch/x86/kvm/vmx/vmx.c static inline bool cpu_need_virtualize_apic_accesses(struct kvm_vcpu *vcpu) kvm_vcpu 751 arch/x86/kvm/vmx/vmx.c void update_exception_bitmap(struct kvm_vcpu *vcpu) kvm_vcpu 788 arch/x86/kvm/vmx/vmx.c static bool msr_write_intercepted(struct kvm_vcpu *vcpu, u32 msr) kvm_vcpu 1107 arch/x86/kvm/vmx/vmx.c void vmx_prepare_switch_to_guest(struct kvm_vcpu *vcpu) kvm_vcpu 1233 arch/x86/kvm/vmx/vmx.c static void vmx_vcpu_pi_load(struct kvm_vcpu *vcpu, int cpu) kvm_vcpu 1289 arch/x86/kvm/vmx/vmx.c void vmx_vcpu_load_vmcs(struct kvm_vcpu *vcpu, int cpu, kvm_vcpu 1365 arch/x86/kvm/vmx/vmx.c void vmx_vcpu_load(struct kvm_vcpu *vcpu, int cpu) kvm_vcpu 1376 arch/x86/kvm/vmx/vmx.c static void vmx_vcpu_pi_put(struct kvm_vcpu *vcpu) kvm_vcpu 1390 arch/x86/kvm/vmx/vmx.c static void vmx_vcpu_put(struct kvm_vcpu *vcpu) kvm_vcpu 1397 arch/x86/kvm/vmx/vmx.c static bool emulation_required(struct kvm_vcpu *vcpu) kvm_vcpu 1402 arch/x86/kvm/vmx/vmx.c static void vmx_decache_cr0_guest_bits(struct kvm_vcpu *vcpu); kvm_vcpu 1404 arch/x86/kvm/vmx/vmx.c unsigned long vmx_get_rflags(struct kvm_vcpu *vcpu) kvm_vcpu 1421 arch/x86/kvm/vmx/vmx.c void vmx_set_rflags(struct kvm_vcpu *vcpu, unsigned long rflags) kvm_vcpu 1437 arch/x86/kvm/vmx/vmx.c u32 vmx_get_interrupt_shadow(struct kvm_vcpu *vcpu) kvm_vcpu 1450 arch/x86/kvm/vmx/vmx.c void vmx_set_interrupt_shadow(struct kvm_vcpu *vcpu, int mask) kvm_vcpu 1466 arch/x86/kvm/vmx/vmx.c static int vmx_rtit_ctl_check(struct kvm_vcpu *vcpu, u64 data) kvm_vcpu 1538 arch/x86/kvm/vmx/vmx.c static int skip_emulated_instruction(struct kvm_vcpu *vcpu) kvm_vcpu 1566 arch/x86/kvm/vmx/vmx.c static void vmx_clear_hlt(struct kvm_vcpu *vcpu) kvm_vcpu 1579 arch/x86/kvm/vmx/vmx.c static void vmx_queue_exception(struct kvm_vcpu *vcpu) kvm_vcpu 1679 arch/x86/kvm/vmx/vmx.c static u64 vmx_read_l1_tsc_offset(struct kvm_vcpu *vcpu) kvm_vcpu 1690 arch/x86/kvm/vmx/vmx.c static u64 vmx_write_l1_tsc_offset(struct kvm_vcpu *vcpu, u64 offset) kvm_vcpu 1718 arch/x86/kvm/vmx/vmx.c bool nested_vmx_allowed(struct kvm_vcpu *vcpu) kvm_vcpu 1723 arch/x86/kvm/vmx/vmx.c static inline bool vmx_feature_control_msr_valid(struct kvm_vcpu *vcpu, kvm_vcpu 1750 arch/x86/kvm/vmx/vmx.c static int vmx_get_msr(struct kvm_vcpu *vcpu, struct msr_data *msr_info) kvm_vcpu 1890 arch/x86/kvm/vmx/vmx.c static int vmx_set_msr(struct kvm_vcpu *vcpu, struct msr_data *msr_info) kvm_vcpu 2164 arch/x86/kvm/vmx/vmx.c static void vmx_cache_reg(struct kvm_vcpu *vcpu, enum kvm_reg reg) kvm_vcpu 2624 arch/x86/kvm/vmx/vmx.c static void fix_pmode_seg(struct kvm_vcpu *vcpu, int seg, kvm_vcpu 2643 arch/x86/kvm/vmx/vmx.c static void enter_pmode(struct kvm_vcpu *vcpu) kvm_vcpu 2716 arch/x86/kvm/vmx/vmx.c static void enter_rmode(struct kvm_vcpu *vcpu) kvm_vcpu 2765 arch/x86/kvm/vmx/vmx.c void vmx_set_efer(struct kvm_vcpu *vcpu, u64 efer) kvm_vcpu 2787 arch/x86/kvm/vmx/vmx.c static void enter_lmode(struct kvm_vcpu *vcpu) kvm_vcpu 2804 arch/x86/kvm/vmx/vmx.c static void exit_lmode(struct kvm_vcpu *vcpu) kvm_vcpu 2812 arch/x86/kvm/vmx/vmx.c static void vmx_flush_tlb_gva(struct kvm_vcpu *vcpu, gva_t addr) kvm_vcpu 2826 arch/x86/kvm/vmx/vmx.c static void vmx_decache_cr0_guest_bits(struct kvm_vcpu *vcpu) kvm_vcpu 2834 arch/x86/kvm/vmx/vmx.c static void vmx_decache_cr3(struct kvm_vcpu *vcpu) kvm_vcpu 2841 arch/x86/kvm/vmx/vmx.c static void vmx_decache_cr4_guest_bits(struct kvm_vcpu *vcpu) kvm_vcpu 2849 arch/x86/kvm/vmx/vmx.c static void ept_load_pdptrs(struct kvm_vcpu *vcpu) kvm_vcpu 2865 arch/x86/kvm/vmx/vmx.c void ept_save_pdptrs(struct kvm_vcpu *vcpu) kvm_vcpu 2884 arch/x86/kvm/vmx/vmx.c struct kvm_vcpu *vcpu) kvm_vcpu 2908 arch/x86/kvm/vmx/vmx.c void vmx_set_cr0(struct kvm_vcpu *vcpu, unsigned long cr0) kvm_vcpu 2946 arch/x86/kvm/vmx/vmx.c static int get_ept_level(struct kvm_vcpu *vcpu) kvm_vcpu 2956 arch/x86/kvm/vmx/vmx.c u64 construct_eptp(struct kvm_vcpu *vcpu, unsigned long root_hpa) kvm_vcpu 2970 arch/x86/kvm/vmx/vmx.c void vmx_set_cr3(struct kvm_vcpu *vcpu, unsigned long cr3) kvm_vcpu 3004 arch/x86/kvm/vmx/vmx.c int vmx_set_cr4(struct kvm_vcpu *vcpu, unsigned long cr4) kvm_vcpu 3079 arch/x86/kvm/vmx/vmx.c void vmx_get_segment(struct kvm_vcpu *vcpu, struct kvm_segment *var, int seg) kvm_vcpu 3115 arch/x86/kvm/vmx/vmx.c static u64 vmx_get_segment_base(struct kvm_vcpu *vcpu, int seg) kvm_vcpu 3126 arch/x86/kvm/vmx/vmx.c int vmx_get_cpl(struct kvm_vcpu *vcpu) kvm_vcpu 3158 arch/x86/kvm/vmx/vmx.c void vmx_set_segment(struct kvm_vcpu *vcpu, struct kvm_segment *var, int seg) kvm_vcpu 3198 arch/x86/kvm/vmx/vmx.c static void vmx_get_cs_db_l_bits(struct kvm_vcpu *vcpu, int *db, int *l) kvm_vcpu 3206 arch/x86/kvm/vmx/vmx.c static void vmx_get_idt(struct kvm_vcpu *vcpu, struct desc_ptr *dt) kvm_vcpu 3212 arch/x86/kvm/vmx/vmx.c static void vmx_set_idt(struct kvm_vcpu *vcpu, struct desc_ptr *dt) kvm_vcpu 3218 arch/x86/kvm/vmx/vmx.c static void vmx_get_gdt(struct kvm_vcpu *vcpu, struct desc_ptr *dt) kvm_vcpu 3224 arch/x86/kvm/vmx/vmx.c static void vmx_set_gdt(struct kvm_vcpu *vcpu, struct desc_ptr *dt) kvm_vcpu 3230 arch/x86/kvm/vmx/vmx.c static bool rmode_segment_valid(struct kvm_vcpu *vcpu, int seg) kvm_vcpu 3251 arch/x86/kvm/vmx/vmx.c static bool code_segment_valid(struct kvm_vcpu *vcpu) kvm_vcpu 3279 arch/x86/kvm/vmx/vmx.c static bool stack_segment_valid(struct kvm_vcpu *vcpu) kvm_vcpu 3301 arch/x86/kvm/vmx/vmx.c static bool data_segment_valid(struct kvm_vcpu *vcpu, int seg) kvm_vcpu 3326 arch/x86/kvm/vmx/vmx.c static bool tr_valid(struct kvm_vcpu *vcpu) kvm_vcpu 3344 arch/x86/kvm/vmx/vmx.c static bool ldtr_valid(struct kvm_vcpu *vcpu) kvm_vcpu 3362 arch/x86/kvm/vmx/vmx.c static bool cs_ss_rpl_check(struct kvm_vcpu *vcpu) kvm_vcpu 3378 arch/x86/kvm/vmx/vmx.c static bool guest_state_valid(struct kvm_vcpu *vcpu) kvm_vcpu 3657 arch/x86/kvm/vmx/vmx.c static u8 vmx_msr_bitmap_mode(struct kvm_vcpu *vcpu) kvm_vcpu 3697 arch/x86/kvm/vmx/vmx.c void vmx_update_msr_bitmap(struct kvm_vcpu *vcpu) kvm_vcpu 3735 arch/x86/kvm/vmx/vmx.c static bool vmx_get_enable_apicv(struct kvm_vcpu *vcpu) kvm_vcpu 3740 arch/x86/kvm/vmx/vmx.c static bool vmx_guest_apic_has_interrupt(struct kvm_vcpu *vcpu) kvm_vcpu 3760 arch/x86/kvm/vmx/vmx.c static inline bool kvm_vcpu_trigger_posted_interrupt(struct kvm_vcpu *vcpu, kvm_vcpu 3799 arch/x86/kvm/vmx/vmx.c static int vmx_deliver_nested_posted_interrupt(struct kvm_vcpu *vcpu, kvm_vcpu 3826 arch/x86/kvm/vmx/vmx.c static int vmx_deliver_posted_interrupt(struct kvm_vcpu *vcpu, int vector) kvm_vcpu 3941 arch/x86/kvm/vmx/vmx.c static void vmx_refresh_apicv_exec_ctrl(struct kvm_vcpu *vcpu) kvm_vcpu 3990 arch/x86/kvm/vmx/vmx.c struct kvm_vcpu *vcpu = &vmx->vcpu; kvm_vcpu 4254 arch/x86/kvm/vmx/vmx.c static void vmx_vcpu_reset(struct kvm_vcpu *vcpu, bool init_event) kvm_vcpu 4353 arch/x86/kvm/vmx/vmx.c static void enable_irq_window(struct kvm_vcpu *vcpu) kvm_vcpu 4358 arch/x86/kvm/vmx/vmx.c static void enable_nmi_window(struct kvm_vcpu *vcpu) kvm_vcpu 4369 arch/x86/kvm/vmx/vmx.c static void vmx_inject_irq(struct kvm_vcpu *vcpu) kvm_vcpu 4397 arch/x86/kvm/vmx/vmx.c static void vmx_inject_nmi(struct kvm_vcpu *vcpu) kvm_vcpu 4428 arch/x86/kvm/vmx/vmx.c bool vmx_get_nmi_mask(struct kvm_vcpu *vcpu) kvm_vcpu 4442 arch/x86/kvm/vmx/vmx.c void vmx_set_nmi_mask(struct kvm_vcpu *vcpu, bool masked) kvm_vcpu 4462 arch/x86/kvm/vmx/vmx.c static int vmx_nmi_allowed(struct kvm_vcpu *vcpu) kvm_vcpu 4476 arch/x86/kvm/vmx/vmx.c static int vmx_interrupt_allowed(struct kvm_vcpu *vcpu) kvm_vcpu 4510 arch/x86/kvm/vmx/vmx.c static bool rmode_exception(struct kvm_vcpu *vcpu, int vec) kvm_vcpu 4542 arch/x86/kvm/vmx/vmx.c static int handle_rmode_exception(struct kvm_vcpu *vcpu, kvm_vcpu 4588 arch/x86/kvm/vmx/vmx.c static int handle_machine_check(struct kvm_vcpu *vcpu) kvm_vcpu 4594 arch/x86/kvm/vmx/vmx.c static int handle_exception_nmi(struct kvm_vcpu *vcpu) kvm_vcpu 4699 arch/x86/kvm/vmx/vmx.c static int handle_external_interrupt(struct kvm_vcpu *vcpu) kvm_vcpu 4705 arch/x86/kvm/vmx/vmx.c static int handle_triple_fault(struct kvm_vcpu *vcpu) kvm_vcpu 4712 arch/x86/kvm/vmx/vmx.c static int handle_io(struct kvm_vcpu *vcpu) kvm_vcpu 4734 arch/x86/kvm/vmx/vmx.c vmx_patch_hypercall(struct kvm_vcpu *vcpu, unsigned char *hypercall) kvm_vcpu 4745 arch/x86/kvm/vmx/vmx.c static int handle_set_cr0(struct kvm_vcpu *vcpu, unsigned long val) kvm_vcpu 4778 arch/x86/kvm/vmx/vmx.c static int handle_set_cr4(struct kvm_vcpu *vcpu, unsigned long val) kvm_vcpu 4795 arch/x86/kvm/vmx/vmx.c static int handle_desc(struct kvm_vcpu *vcpu) kvm_vcpu 4801 arch/x86/kvm/vmx/vmx.c static int handle_cr(struct kvm_vcpu *vcpu) kvm_vcpu 4881 arch/x86/kvm/vmx/vmx.c static int handle_dr(struct kvm_vcpu *vcpu) kvm_vcpu 4944 arch/x86/kvm/vmx/vmx.c static u64 vmx_get_dr6(struct kvm_vcpu *vcpu) kvm_vcpu 4949 arch/x86/kvm/vmx/vmx.c static void vmx_set_dr6(struct kvm_vcpu *vcpu, unsigned long val) kvm_vcpu 4953 arch/x86/kvm/vmx/vmx.c static void vmx_sync_dirty_debug_regs(struct kvm_vcpu *vcpu) kvm_vcpu 4966 arch/x86/kvm/vmx/vmx.c static void vmx_set_dr7(struct kvm_vcpu *vcpu, unsigned long val) kvm_vcpu 4971 arch/x86/kvm/vmx/vmx.c static int handle_cpuid(struct kvm_vcpu *vcpu) kvm_vcpu 4976 arch/x86/kvm/vmx/vmx.c static int handle_rdmsr(struct kvm_vcpu *vcpu) kvm_vcpu 4981 arch/x86/kvm/vmx/vmx.c static int handle_wrmsr(struct kvm_vcpu *vcpu) kvm_vcpu 4986 arch/x86/kvm/vmx/vmx.c static int handle_tpr_below_threshold(struct kvm_vcpu *vcpu) kvm_vcpu 4992 arch/x86/kvm/vmx/vmx.c static int handle_interrupt_window(struct kvm_vcpu *vcpu) kvm_vcpu 5002 arch/x86/kvm/vmx/vmx.c static int handle_halt(struct kvm_vcpu *vcpu) kvm_vcpu 5007 arch/x86/kvm/vmx/vmx.c static int handle_vmcall(struct kvm_vcpu *vcpu) kvm_vcpu 5012 arch/x86/kvm/vmx/vmx.c static int handle_invd(struct kvm_vcpu *vcpu) kvm_vcpu 5017 arch/x86/kvm/vmx/vmx.c static int handle_invlpg(struct kvm_vcpu *vcpu) kvm_vcpu 5025 arch/x86/kvm/vmx/vmx.c static int handle_rdpmc(struct kvm_vcpu *vcpu) kvm_vcpu 5033 arch/x86/kvm/vmx/vmx.c static int handle_wbinvd(struct kvm_vcpu *vcpu) kvm_vcpu 5038 arch/x86/kvm/vmx/vmx.c static int handle_xsetbv(struct kvm_vcpu *vcpu) kvm_vcpu 5048 arch/x86/kvm/vmx/vmx.c static int handle_apic_access(struct kvm_vcpu *vcpu) kvm_vcpu 5070 arch/x86/kvm/vmx/vmx.c static int handle_apic_eoi_induced(struct kvm_vcpu *vcpu) kvm_vcpu 5080 arch/x86/kvm/vmx/vmx.c static int handle_apic_write(struct kvm_vcpu *vcpu) kvm_vcpu 5090 arch/x86/kvm/vmx/vmx.c static int handle_task_switch(struct kvm_vcpu *vcpu) kvm_vcpu 5147 arch/x86/kvm/vmx/vmx.c static int handle_ept_violation(struct kvm_vcpu *vcpu) kvm_vcpu 5191 arch/x86/kvm/vmx/vmx.c static int handle_ept_misconfig(struct kvm_vcpu *vcpu) kvm_vcpu 5209 arch/x86/kvm/vmx/vmx.c static int handle_nmi_window(struct kvm_vcpu *vcpu) kvm_vcpu 5219 arch/x86/kvm/vmx/vmx.c static int handle_invalid_guest_state(struct kvm_vcpu *vcpu) kvm_vcpu 5273 arch/x86/kvm/vmx/vmx.c static void grow_ple_window(struct kvm_vcpu *vcpu) kvm_vcpu 5289 arch/x86/kvm/vmx/vmx.c static void shrink_ple_window(struct kvm_vcpu *vcpu) kvm_vcpu 5310 arch/x86/kvm/vmx/vmx.c struct kvm_vcpu *vcpu; kvm_vcpu 5341 arch/x86/kvm/vmx/vmx.c static int handle_pause(struct kvm_vcpu *vcpu) kvm_vcpu 5356 arch/x86/kvm/vmx/vmx.c static int handle_nop(struct kvm_vcpu *vcpu) kvm_vcpu 5361 arch/x86/kvm/vmx/vmx.c static int handle_mwait(struct kvm_vcpu *vcpu) kvm_vcpu 5367 arch/x86/kvm/vmx/vmx.c static int handle_invalid_op(struct kvm_vcpu *vcpu) kvm_vcpu 5373 arch/x86/kvm/vmx/vmx.c static int handle_monitor_trap(struct kvm_vcpu *vcpu) kvm_vcpu 5378 arch/x86/kvm/vmx/vmx.c static int handle_monitor(struct kvm_vcpu *vcpu) kvm_vcpu 5384 arch/x86/kvm/vmx/vmx.c static int handle_invpcid(struct kvm_vcpu *vcpu) kvm_vcpu 5484 arch/x86/kvm/vmx/vmx.c static int handle_pml_full(struct kvm_vcpu *vcpu) kvm_vcpu 5509 arch/x86/kvm/vmx/vmx.c static int handle_preemption_timer(struct kvm_vcpu *vcpu) kvm_vcpu 5524 arch/x86/kvm/vmx/vmx.c static int handle_vmx_instruction(struct kvm_vcpu *vcpu) kvm_vcpu 5530 arch/x86/kvm/vmx/vmx.c static int handle_encls(struct kvm_vcpu *vcpu) kvm_vcpu 5546 arch/x86/kvm/vmx/vmx.c static int (*kvm_vmx_exit_handlers[])(struct kvm_vcpu *vcpu) = { kvm_vcpu 5602 arch/x86/kvm/vmx/vmx.c static void vmx_get_exit_info(struct kvm_vcpu *vcpu, u64 *info1, u64 *info2) kvm_vcpu 5616 arch/x86/kvm/vmx/vmx.c static void vmx_flush_pml_buffer(struct kvm_vcpu *vcpu) kvm_vcpu 5654 arch/x86/kvm/vmx/vmx.c struct kvm_vcpu *vcpu; kvm_vcpu 5844 arch/x86/kvm/vmx/vmx.c static int vmx_handle_exit(struct kvm_vcpu *vcpu) kvm_vcpu 5956 arch/x86/kvm/vmx/vmx.c static void vmx_l1d_flush(struct kvm_vcpu *vcpu) kvm_vcpu 6016 arch/x86/kvm/vmx/vmx.c static void update_cr8_intercept(struct kvm_vcpu *vcpu, int tpr, int irr) kvm_vcpu 6032 arch/x86/kvm/vmx/vmx.c void vmx_set_virtual_apic_mode(struct kvm_vcpu *vcpu) kvm_vcpu 6077 arch/x86/kvm/vmx/vmx.c static void vmx_set_apic_access_page_addr(struct kvm_vcpu *vcpu, hpa_t hpa) kvm_vcpu 6085 arch/x86/kvm/vmx/vmx.c static void vmx_hwapic_isr_update(struct kvm_vcpu *vcpu, int max_isr) kvm_vcpu 6119 arch/x86/kvm/vmx/vmx.c static void vmx_hwapic_irr_update(struct kvm_vcpu *vcpu, int max_irr) kvm_vcpu 6133 arch/x86/kvm/vmx/vmx.c static int vmx_sync_pir_to_irr(struct kvm_vcpu *vcpu) kvm_vcpu 6171 arch/x86/kvm/vmx/vmx.c static bool vmx_dy_apicv_has_pending_interrupt(struct kvm_vcpu *vcpu) kvm_vcpu 6179 arch/x86/kvm/vmx/vmx.c static void vmx_load_eoi_exitmap(struct kvm_vcpu *vcpu, u64 *eoi_exit_bitmap) kvm_vcpu 6190 arch/x86/kvm/vmx/vmx.c static void vmx_apicv_post_state_restore(struct kvm_vcpu *vcpu) kvm_vcpu 6218 arch/x86/kvm/vmx/vmx.c static void handle_external_interrupt_irqoff(struct kvm_vcpu *vcpu) kvm_vcpu 6264 arch/x86/kvm/vmx/vmx.c static void vmx_handle_exit_irqoff(struct kvm_vcpu *vcpu) kvm_vcpu 6341 arch/x86/kvm/vmx/vmx.c static void __vmx_complete_interrupts(struct kvm_vcpu *vcpu, kvm_vcpu 6402 arch/x86/kvm/vmx/vmx.c static void vmx_cancel_injection(struct kvm_vcpu *vcpu) kvm_vcpu 6447 arch/x86/kvm/vmx/vmx.c static void vmx_update_hv_timer(struct kvm_vcpu *vcpu) kvm_vcpu 6483 arch/x86/kvm/vmx/vmx.c static void vmx_vcpu_run(struct kvm_vcpu *vcpu) kvm_vcpu 6658 arch/x86/kvm/vmx/vmx.c static void vmx_free_vcpu(struct kvm_vcpu *vcpu) kvm_vcpu 6674 arch/x86/kvm/vmx/vmx.c static struct kvm_vcpu *vmx_create_vcpu(struct kvm *kvm, unsigned int id) kvm_vcpu 6863 arch/x86/kvm/vmx/vmx.c static u64 vmx_get_mt_mask(struct kvm_vcpu *vcpu, gfn_t gfn, bool is_mmio) kvm_vcpu 6938 arch/x86/kvm/vmx/vmx.c static void nested_vmx_cr_fixed1_bits_update(struct kvm_vcpu *vcpu) kvm_vcpu 6977 arch/x86/kvm/vmx/vmx.c static void nested_vmx_entry_exit_ctls_update(struct kvm_vcpu *vcpu) kvm_vcpu 6994 arch/x86/kvm/vmx/vmx.c static void update_intel_pt_cfg(struct kvm_vcpu *vcpu) kvm_vcpu 7063 arch/x86/kvm/vmx/vmx.c static void vmx_cpuid_update(struct kvm_vcpu *vcpu) kvm_vcpu 7095 arch/x86/kvm/vmx/vmx.c static void vmx_request_immediate_exit(struct kvm_vcpu *vcpu) kvm_vcpu 7100 arch/x86/kvm/vmx/vmx.c static int vmx_check_intercept_io(struct kvm_vcpu *vcpu, kvm_vcpu 7134 arch/x86/kvm/vmx/vmx.c static int vmx_check_intercept(struct kvm_vcpu *vcpu, kvm_vcpu 7201 arch/x86/kvm/vmx/vmx.c static int vmx_set_hv_timer(struct kvm_vcpu *vcpu, u64 guest_deadline_tsc, kvm_vcpu 7245 arch/x86/kvm/vmx/vmx.c static void vmx_cancel_hv_timer(struct kvm_vcpu *vcpu) kvm_vcpu 7251 arch/x86/kvm/vmx/vmx.c static void vmx_sched_in(struct kvm_vcpu *vcpu, int cpu) kvm_vcpu 7275 arch/x86/kvm/vmx/vmx.c static int vmx_write_pml_buffer(struct kvm_vcpu *vcpu) kvm_vcpu 7318 arch/x86/kvm/vmx/vmx.c static void __pi_post_block(struct kvm_vcpu *vcpu) kvm_vcpu 7362 arch/x86/kvm/vmx/vmx.c static int pi_pre_block(struct kvm_vcpu *vcpu) kvm_vcpu 7419 arch/x86/kvm/vmx/vmx.c static int vmx_pre_block(struct kvm_vcpu *vcpu) kvm_vcpu 7430 arch/x86/kvm/vmx/vmx.c static void pi_post_block(struct kvm_vcpu *vcpu) kvm_vcpu 7441 arch/x86/kvm/vmx/vmx.c static void vmx_post_block(struct kvm_vcpu *vcpu) kvm_vcpu 7464 arch/x86/kvm/vmx/vmx.c struct kvm_vcpu *vcpu; kvm_vcpu 7543 arch/x86/kvm/vmx/vmx.c static void vmx_setup_mce(struct kvm_vcpu *vcpu) kvm_vcpu 7553 arch/x86/kvm/vmx/vmx.c static int vmx_smi_allowed(struct kvm_vcpu *vcpu) kvm_vcpu 7561 arch/x86/kvm/vmx/vmx.c static int vmx_pre_enter_smm(struct kvm_vcpu *vcpu, char *smstate) kvm_vcpu 7575 arch/x86/kvm/vmx/vmx.c static int vmx_pre_leave_smm(struct kvm_vcpu *vcpu, const char *smstate) kvm_vcpu 7595 arch/x86/kvm/vmx/vmx.c static int enable_smi_window(struct kvm_vcpu *vcpu) kvm_vcpu 7600 arch/x86/kvm/vmx/vmx.c static bool vmx_need_emulation_on_page_fault(struct kvm_vcpu *vcpu) kvm_vcpu 7605 arch/x86/kvm/vmx/vmx.c static bool vmx_apic_init_signal_blocked(struct kvm_vcpu *vcpu) kvm_vcpu 189 arch/x86/kvm/vmx/vmx.h struct kvm_vcpu vcpu; kvm_vcpu 306 arch/x86/kvm/vmx/vmx.h bool nested_vmx_allowed(struct kvm_vcpu *vcpu); kvm_vcpu 307 arch/x86/kvm/vmx/vmx.h void vmx_vcpu_load_vmcs(struct kvm_vcpu *vcpu, int cpu, kvm_vcpu 309 arch/x86/kvm/vmx/vmx.h void vmx_vcpu_load(struct kvm_vcpu *vcpu, int cpu); kvm_vcpu 313 arch/x86/kvm/vmx/vmx.h void vmx_prepare_switch_to_guest(struct kvm_vcpu *vcpu); kvm_vcpu 316 arch/x86/kvm/vmx/vmx.h int vmx_get_cpl(struct kvm_vcpu *vcpu); kvm_vcpu 317 arch/x86/kvm/vmx/vmx.h unsigned long vmx_get_rflags(struct kvm_vcpu *vcpu); kvm_vcpu 318 arch/x86/kvm/vmx/vmx.h void vmx_set_rflags(struct kvm_vcpu *vcpu, unsigned long rflags); kvm_vcpu 319 arch/x86/kvm/vmx/vmx.h u32 vmx_get_interrupt_shadow(struct kvm_vcpu *vcpu); kvm_vcpu 320 arch/x86/kvm/vmx/vmx.h void vmx_set_interrupt_shadow(struct kvm_vcpu *vcpu, int mask); kvm_vcpu 321 arch/x86/kvm/vmx/vmx.h void vmx_set_efer(struct kvm_vcpu *vcpu, u64 efer); kvm_vcpu 322 arch/x86/kvm/vmx/vmx.h void vmx_set_cr0(struct kvm_vcpu *vcpu, unsigned long cr0); kvm_vcpu 323 arch/x86/kvm/vmx/vmx.h void vmx_set_cr3(struct kvm_vcpu *vcpu, unsigned long cr3); kvm_vcpu 324 arch/x86/kvm/vmx/vmx.h int vmx_set_cr4(struct kvm_vcpu *vcpu, unsigned long cr4); kvm_vcpu 326 arch/x86/kvm/vmx/vmx.h void ept_save_pdptrs(struct kvm_vcpu *vcpu); kvm_vcpu 327 arch/x86/kvm/vmx/vmx.h void vmx_get_segment(struct kvm_vcpu *vcpu, struct kvm_segment *var, int seg); kvm_vcpu 328 arch/x86/kvm/vmx/vmx.h void vmx_set_segment(struct kvm_vcpu *vcpu, struct kvm_segment *var, int seg); kvm_vcpu 329 arch/x86/kvm/vmx/vmx.h u64 construct_eptp(struct kvm_vcpu *vcpu, unsigned long root_hpa); kvm_vcpu 330 arch/x86/kvm/vmx/vmx.h void update_exception_bitmap(struct kvm_vcpu *vcpu); kvm_vcpu 331 arch/x86/kvm/vmx/vmx.h void vmx_update_msr_bitmap(struct kvm_vcpu *vcpu); kvm_vcpu 332 arch/x86/kvm/vmx/vmx.h bool vmx_get_nmi_mask(struct kvm_vcpu *vcpu); kvm_vcpu 333 arch/x86/kvm/vmx/vmx.h void vmx_set_nmi_mask(struct kvm_vcpu *vcpu, bool masked); kvm_vcpu 334 arch/x86/kvm/vmx/vmx.h void vmx_set_virtual_apic_mode(struct kvm_vcpu *vcpu); kvm_vcpu 466 arch/x86/kvm/vmx/vmx.h static inline struct vcpu_vmx *to_vmx(struct kvm_vcpu *vcpu) kvm_vcpu 471 arch/x86/kvm/vmx/vmx.h static inline struct pi_desc *vcpu_to_pi_desc(struct kvm_vcpu *vcpu) kvm_vcpu 489 arch/x86/kvm/vmx/vmx.h u64 construct_eptp(struct kvm_vcpu *vcpu, unsigned long root_hpa); kvm_vcpu 491 arch/x86/kvm/vmx/vmx.h static inline void __vmx_flush_tlb(struct kvm_vcpu *vcpu, int vpid, kvm_vcpu 504 arch/x86/kvm/vmx/vmx.h static inline void vmx_flush_tlb(struct kvm_vcpu *vcpu, bool invalidate_gpa) kvm_vcpu 82 arch/x86/kvm/x86.c container_of(ctxt, struct kvm_vcpu, arch.emulate_ctxt) kvm_vcpu 98 arch/x86/kvm/x86.c #define VCPU_STAT(x, ...) offsetof(struct kvm_vcpu, stat.x), KVM_STAT_VCPU, ## __VA_ARGS__ kvm_vcpu 103 arch/x86/kvm/x86.c static void update_cr8_intercept(struct kvm_vcpu *vcpu); kvm_vcpu 104 arch/x86/kvm/x86.c static void process_nmi(struct kvm_vcpu *vcpu); kvm_vcpu 105 arch/x86/kvm/x86.c static void enter_smm(struct kvm_vcpu *vcpu); kvm_vcpu 106 arch/x86/kvm/x86.c static void __kvm_set_rflags(struct kvm_vcpu *vcpu, unsigned long rflags); kvm_vcpu 107 arch/x86/kvm/x86.c static void store_regs(struct kvm_vcpu *vcpu); kvm_vcpu 108 arch/x86/kvm/x86.c static int sync_regs(struct kvm_vcpu *vcpu); kvm_vcpu 231 arch/x86/kvm/x86.c static inline void kvm_async_pf_hash_reset(struct kvm_vcpu *vcpu) kvm_vcpu 331 arch/x86/kvm/x86.c u64 kvm_get_apic_base(struct kvm_vcpu *vcpu) kvm_vcpu 337 arch/x86/kvm/x86.c enum lapic_mode kvm_get_apic_mode(struct kvm_vcpu *vcpu) kvm_vcpu 343 arch/x86/kvm/x86.c int kvm_set_apic_base(struct kvm_vcpu *vcpu, struct msr_data *msr_info) kvm_vcpu 417 arch/x86/kvm/x86.c void kvm_deliver_exception_payload(struct kvm_vcpu *vcpu) kvm_vcpu 467 arch/x86/kvm/x86.c static void kvm_multiple_exception(struct kvm_vcpu *vcpu, kvm_vcpu 555 arch/x86/kvm/x86.c void kvm_queue_exception(struct kvm_vcpu *vcpu, unsigned nr) kvm_vcpu 561 arch/x86/kvm/x86.c void kvm_requeue_exception(struct kvm_vcpu *vcpu, unsigned nr) kvm_vcpu 567 arch/x86/kvm/x86.c static void kvm_queue_exception_p(struct kvm_vcpu *vcpu, unsigned nr, kvm_vcpu 573 arch/x86/kvm/x86.c static void kvm_queue_exception_e_p(struct kvm_vcpu *vcpu, unsigned nr, kvm_vcpu 580 arch/x86/kvm/x86.c int kvm_complete_insn_gp(struct kvm_vcpu *vcpu, int err) kvm_vcpu 591 arch/x86/kvm/x86.c void kvm_inject_page_fault(struct kvm_vcpu *vcpu, struct x86_exception *fault) kvm_vcpu 606 arch/x86/kvm/x86.c static bool kvm_propagate_fault(struct kvm_vcpu *vcpu, struct x86_exception *fault) kvm_vcpu 616 arch/x86/kvm/x86.c void kvm_inject_nmi(struct kvm_vcpu *vcpu) kvm_vcpu 623 arch/x86/kvm/x86.c void kvm_queue_exception_e(struct kvm_vcpu *vcpu, unsigned nr, u32 error_code) kvm_vcpu 629 arch/x86/kvm/x86.c void kvm_requeue_exception_e(struct kvm_vcpu *vcpu, unsigned nr, u32 error_code) kvm_vcpu 639 arch/x86/kvm/x86.c bool kvm_require_cpl(struct kvm_vcpu *vcpu, int required_cpl) kvm_vcpu 648 arch/x86/kvm/x86.c bool kvm_require_dr(struct kvm_vcpu *vcpu, int dr) kvm_vcpu 663 arch/x86/kvm/x86.c int kvm_read_guest_page_mmu(struct kvm_vcpu *vcpu, struct kvm_mmu *mmu, kvm_vcpu 682 arch/x86/kvm/x86.c static int kvm_read_nested_guest_page(struct kvm_vcpu *vcpu, gfn_t gfn, kvm_vcpu 689 arch/x86/kvm/x86.c static inline u64 pdptr_rsvd_bits(struct kvm_vcpu *vcpu) kvm_vcpu 698 arch/x86/kvm/x86.c int load_pdptrs(struct kvm_vcpu *vcpu, struct kvm_mmu *mmu, unsigned long cr3) kvm_vcpu 733 arch/x86/kvm/x86.c bool pdptrs_changed(struct kvm_vcpu *vcpu) kvm_vcpu 761 arch/x86/kvm/x86.c int kvm_set_cr0(struct kvm_vcpu *vcpu, unsigned long cr0) kvm_vcpu 820 arch/x86/kvm/x86.c void kvm_lmsw(struct kvm_vcpu *vcpu, unsigned long msw) kvm_vcpu 826 arch/x86/kvm/x86.c void kvm_load_guest_xcr0(struct kvm_vcpu *vcpu) kvm_vcpu 844 arch/x86/kvm/x86.c void kvm_put_guest_xcr0(struct kvm_vcpu *vcpu) kvm_vcpu 862 arch/x86/kvm/x86.c static int __kvm_set_xcr(struct kvm_vcpu *vcpu, u32 index, u64 xcr) kvm_vcpu 902 arch/x86/kvm/x86.c int kvm_set_xcr(struct kvm_vcpu *vcpu, u32 index, u64 xcr) kvm_vcpu 942 arch/x86/kvm/x86.c static int kvm_valid_cr4(struct kvm_vcpu *vcpu, unsigned long cr4) kvm_vcpu 971 arch/x86/kvm/x86.c int kvm_set_cr4(struct kvm_vcpu *vcpu, unsigned long cr4) kvm_vcpu 1012 arch/x86/kvm/x86.c int kvm_set_cr3(struct kvm_vcpu *vcpu, unsigned long cr3) kvm_vcpu 1047 arch/x86/kvm/x86.c int kvm_set_cr8(struct kvm_vcpu *vcpu, unsigned long cr8) kvm_vcpu 1059 arch/x86/kvm/x86.c unsigned long kvm_get_cr8(struct kvm_vcpu *vcpu) kvm_vcpu 1068 arch/x86/kvm/x86.c static void kvm_update_dr0123(struct kvm_vcpu *vcpu) kvm_vcpu 1079 arch/x86/kvm/x86.c static void kvm_update_dr6(struct kvm_vcpu *vcpu) kvm_vcpu 1085 arch/x86/kvm/x86.c static void kvm_update_dr7(struct kvm_vcpu *vcpu) kvm_vcpu 1099 arch/x86/kvm/x86.c static u64 kvm_dr6_fixed(struct kvm_vcpu *vcpu) kvm_vcpu 1108 arch/x86/kvm/x86.c static int __kvm_set_dr(struct kvm_vcpu *vcpu, int dr, unsigned long val) kvm_vcpu 1139 arch/x86/kvm/x86.c int kvm_set_dr(struct kvm_vcpu *vcpu, int dr, unsigned long val) kvm_vcpu 1149 arch/x86/kvm/x86.c int kvm_get_dr(struct kvm_vcpu *vcpu, int dr, unsigned long *val) kvm_vcpu 1175 arch/x86/kvm/x86.c bool kvm_rdpmc(struct kvm_vcpu *vcpu) kvm_vcpu 1416 arch/x86/kvm/x86.c static int do_get_msr_feature(struct kvm_vcpu *vcpu, unsigned index, u64 *data) kvm_vcpu 1431 arch/x86/kvm/x86.c static bool __kvm_valid_efer(struct kvm_vcpu *vcpu, u64 efer) kvm_vcpu 1449 arch/x86/kvm/x86.c bool kvm_valid_efer(struct kvm_vcpu *vcpu, u64 efer) kvm_vcpu 1458 arch/x86/kvm/x86.c static int set_efer(struct kvm_vcpu *vcpu, struct msr_data *msr_info) kvm_vcpu 1499 arch/x86/kvm/x86.c static int __kvm_set_msr(struct kvm_vcpu *vcpu, u32 index, u64 data, kvm_vcpu 1543 arch/x86/kvm/x86.c static int __kvm_get_msr(struct kvm_vcpu *vcpu, u32 index, u64 *data, kvm_vcpu 1558 arch/x86/kvm/x86.c int kvm_get_msr(struct kvm_vcpu *vcpu, u32 index, u64 *data) kvm_vcpu 1564 arch/x86/kvm/x86.c int kvm_set_msr(struct kvm_vcpu *vcpu, u32 index, u64 data) kvm_vcpu 1570 arch/x86/kvm/x86.c int kvm_emulate_rdmsr(struct kvm_vcpu *vcpu) kvm_vcpu 1589 arch/x86/kvm/x86.c int kvm_emulate_wrmsr(struct kvm_vcpu *vcpu) kvm_vcpu 1608 arch/x86/kvm/x86.c static int do_get_msr(struct kvm_vcpu *vcpu, unsigned index, u64 *data) kvm_vcpu 1613 arch/x86/kvm/x86.c static int do_set_msr(struct kvm_vcpu *vcpu, unsigned index, u64 *data) kvm_vcpu 1662 arch/x86/kvm/x86.c void kvm_set_pending_timer(struct kvm_vcpu *vcpu) kvm_vcpu 1760 arch/x86/kvm/x86.c static int set_tsc_khz(struct kvm_vcpu *vcpu, u32 user_tsc_khz, bool scale) kvm_vcpu 1796 arch/x86/kvm/x86.c static int kvm_set_tsc_khz(struct kvm_vcpu *vcpu, u32 user_tsc_khz) kvm_vcpu 1829 arch/x86/kvm/x86.c static u64 compute_guest_tsc(struct kvm_vcpu *vcpu, s64 kernel_ns) kvm_vcpu 1843 arch/x86/kvm/x86.c static void kvm_track_tsc_matching(struct kvm_vcpu *vcpu) kvm_vcpu 1871 arch/x86/kvm/x86.c static void update_ia32_tsc_adjust_msr(struct kvm_vcpu *vcpu, s64 offset) kvm_vcpu 1892 arch/x86/kvm/x86.c u64 kvm_scale_tsc(struct kvm_vcpu *vcpu, u64 tsc) kvm_vcpu 1904 arch/x86/kvm/x86.c static u64 kvm_compute_tsc_offset(struct kvm_vcpu *vcpu, u64 target_tsc) kvm_vcpu 1913 arch/x86/kvm/x86.c u64 kvm_read_l1_tsc(struct kvm_vcpu *vcpu, u64 host_tsc) kvm_vcpu 1921 arch/x86/kvm/x86.c static void kvm_vcpu_write_tsc_offset(struct kvm_vcpu *vcpu, u64 offset) kvm_vcpu 1939 arch/x86/kvm/x86.c void kvm_write_tsc(struct kvm_vcpu *vcpu, struct msr_data *msr) kvm_vcpu 2044 arch/x86/kvm/x86.c static inline void adjust_tsc_offset_guest(struct kvm_vcpu *vcpu, kvm_vcpu 2051 arch/x86/kvm/x86.c static inline void adjust_tsc_offset_host(struct kvm_vcpu *vcpu, s64 adjustment) kvm_vcpu 2261 arch/x86/kvm/x86.c struct kvm_vcpu *vcpu; kvm_vcpu 2312 arch/x86/kvm/x86.c static void kvm_setup_pvclock_page(struct kvm_vcpu *v) kvm_vcpu 2369 arch/x86/kvm/x86.c static int kvm_guest_time_update(struct kvm_vcpu *v) kvm_vcpu 2482 arch/x86/kvm/x86.c struct kvm_vcpu *vcpu; kvm_vcpu 2490 arch/x86/kvm/x86.c static void kvm_gen_kvmclock_update(struct kvm_vcpu *v) kvm_vcpu 2519 arch/x86/kvm/x86.c static bool can_set_mci_status(struct kvm_vcpu *vcpu) kvm_vcpu 2528 arch/x86/kvm/x86.c static int set_msr_mce(struct kvm_vcpu *vcpu, struct msr_data *msr_info) kvm_vcpu 2578 arch/x86/kvm/x86.c static int xen_hvm_config(struct kvm_vcpu *vcpu, u64 data) kvm_vcpu 2609 arch/x86/kvm/x86.c static int kvm_pv_enable_async_pf(struct kvm_vcpu *vcpu, u64 data) kvm_vcpu 2635 arch/x86/kvm/x86.c static void kvmclock_reset(struct kvm_vcpu *vcpu) kvm_vcpu 2641 arch/x86/kvm/x86.c static void kvm_vcpu_flush_tlb(struct kvm_vcpu *vcpu, bool invalidate_gpa) kvm_vcpu 2647 arch/x86/kvm/x86.c static void record_steal_time(struct kvm_vcpu *vcpu) kvm_vcpu 2692 arch/x86/kvm/x86.c int kvm_set_msr_common(struct kvm_vcpu *vcpu, struct msr_data *msr_info) kvm_vcpu 2952 arch/x86/kvm/x86.c static int get_msr_mce(struct kvm_vcpu *vcpu, u32 msr, u64 *pdata, bool host) kvm_vcpu 2990 arch/x86/kvm/x86.c int kvm_get_msr_common(struct kvm_vcpu *vcpu, struct msr_data *msr_info) kvm_vcpu 3199 arch/x86/kvm/x86.c static int __msr_io(struct kvm_vcpu *vcpu, struct kvm_msrs *msrs, kvm_vcpu 3201 arch/x86/kvm/x86.c int (*do_msr)(struct kvm_vcpu *vcpu, kvm_vcpu 3218 arch/x86/kvm/x86.c static int msr_io(struct kvm_vcpu *vcpu, struct kvm_msrs __user *user_msrs, kvm_vcpu 3219 arch/x86/kvm/x86.c int (*do_msr)(struct kvm_vcpu *vcpu, kvm_vcpu 3488 arch/x86/kvm/x86.c static bool need_emulate_wbinvd(struct kvm_vcpu *vcpu) kvm_vcpu 3493 arch/x86/kvm/x86.c void kvm_arch_vcpu_load(struct kvm_vcpu *vcpu, int cpu) kvm_vcpu 3543 arch/x86/kvm/x86.c static void kvm_steal_time_set_preempted(struct kvm_vcpu *vcpu) kvm_vcpu 3566 arch/x86/kvm/x86.c void kvm_arch_vcpu_put(struct kvm_vcpu *vcpu) kvm_vcpu 3600 arch/x86/kvm/x86.c static int kvm_vcpu_ioctl_get_lapic(struct kvm_vcpu *vcpu, kvm_vcpu 3609 arch/x86/kvm/x86.c static int kvm_vcpu_ioctl_set_lapic(struct kvm_vcpu *vcpu, kvm_vcpu 3622 arch/x86/kvm/x86.c static int kvm_cpu_accept_dm_intr(struct kvm_vcpu *vcpu) kvm_vcpu 3634 arch/x86/kvm/x86.c static int kvm_vcpu_ready_for_interrupt_injection(struct kvm_vcpu *vcpu) kvm_vcpu 3642 arch/x86/kvm/x86.c static int kvm_vcpu_ioctl_interrupt(struct kvm_vcpu *vcpu, kvm_vcpu 3669 arch/x86/kvm/x86.c static int kvm_vcpu_ioctl_nmi(struct kvm_vcpu *vcpu) kvm_vcpu 3676 arch/x86/kvm/x86.c static int kvm_vcpu_ioctl_smi(struct kvm_vcpu *vcpu) kvm_vcpu 3683 arch/x86/kvm/x86.c static int vcpu_ioctl_tpr_access_reporting(struct kvm_vcpu *vcpu, kvm_vcpu 3692 arch/x86/kvm/x86.c static int kvm_vcpu_ioctl_x86_setup_mce(struct kvm_vcpu *vcpu, kvm_vcpu 3717 arch/x86/kvm/x86.c static int kvm_vcpu_ioctl_x86_set_mce(struct kvm_vcpu *vcpu, kvm_vcpu 3765 arch/x86/kvm/x86.c static void kvm_vcpu_ioctl_x86_get_vcpu_events(struct kvm_vcpu *vcpu, kvm_vcpu 3825 arch/x86/kvm/x86.c static void kvm_smm_changed(struct kvm_vcpu *vcpu); kvm_vcpu 3827 arch/x86/kvm/x86.c static int kvm_vcpu_ioctl_x86_set_vcpu_events(struct kvm_vcpu *vcpu, kvm_vcpu 3914 arch/x86/kvm/x86.c static void kvm_vcpu_ioctl_x86_get_debugregs(struct kvm_vcpu *vcpu, kvm_vcpu 3927 arch/x86/kvm/x86.c static int kvm_vcpu_ioctl_x86_set_debugregs(struct kvm_vcpu *vcpu, kvm_vcpu 3950 arch/x86/kvm/x86.c static void fill_xsave(u8 *dest, struct kvm_vcpu *vcpu) kvm_vcpu 3992 arch/x86/kvm/x86.c static void load_xsave(struct kvm_vcpu *vcpu, u8 *src) kvm_vcpu 4034 arch/x86/kvm/x86.c static void kvm_vcpu_ioctl_x86_get_xsave(struct kvm_vcpu *vcpu, kvm_vcpu 4051 arch/x86/kvm/x86.c static int kvm_vcpu_ioctl_x86_set_xsave(struct kvm_vcpu *vcpu, kvm_vcpu 4078 arch/x86/kvm/x86.c static void kvm_vcpu_ioctl_x86_get_xcrs(struct kvm_vcpu *vcpu, kvm_vcpu 4092 arch/x86/kvm/x86.c static int kvm_vcpu_ioctl_x86_set_xcrs(struct kvm_vcpu *vcpu, kvm_vcpu 4121 arch/x86/kvm/x86.c static int kvm_set_guest_paused(struct kvm_vcpu *vcpu) kvm_vcpu 4130 arch/x86/kvm/x86.c static int kvm_vcpu_ioctl_enable_cap(struct kvm_vcpu *vcpu, kvm_vcpu 4176 arch/x86/kvm/x86.c struct kvm_vcpu *vcpu = filp->private_data; kvm_vcpu 4554 arch/x86/kvm/x86.c vm_fault_t kvm_arch_vcpu_fault(struct kvm_vcpu *vcpu, struct vm_fault *vmf) kvm_vcpu 5287 arch/x86/kvm/x86.c static int vcpu_mmio_write(struct kvm_vcpu *vcpu, gpa_t addr, int len, kvm_vcpu 5308 arch/x86/kvm/x86.c static int vcpu_mmio_read(struct kvm_vcpu *vcpu, gpa_t addr, int len, void *v) kvm_vcpu 5330 arch/x86/kvm/x86.c static void kvm_set_segment(struct kvm_vcpu *vcpu, kvm_vcpu 5336 arch/x86/kvm/x86.c void kvm_get_segment(struct kvm_vcpu *vcpu, kvm_vcpu 5342 arch/x86/kvm/x86.c gpa_t translate_nested_gpa(struct kvm_vcpu *vcpu, gpa_t gpa, u32 access, kvm_vcpu 5356 arch/x86/kvm/x86.c gpa_t kvm_mmu_gva_to_gpa_read(struct kvm_vcpu *vcpu, gva_t gva, kvm_vcpu 5363 arch/x86/kvm/x86.c gpa_t kvm_mmu_gva_to_gpa_fetch(struct kvm_vcpu *vcpu, gva_t gva, kvm_vcpu 5371 arch/x86/kvm/x86.c gpa_t kvm_mmu_gva_to_gpa_write(struct kvm_vcpu *vcpu, gva_t gva, kvm_vcpu 5380 arch/x86/kvm/x86.c gpa_t kvm_mmu_gva_to_gpa_system(struct kvm_vcpu *vcpu, gva_t gva, kvm_vcpu 5387 arch/x86/kvm/x86.c struct kvm_vcpu *vcpu, u32 access, kvm_vcpu 5422 arch/x86/kvm/x86.c struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt); kvm_vcpu 5444 arch/x86/kvm/x86.c int kvm_read_guest_virt(struct kvm_vcpu *vcpu, kvm_vcpu 5466 arch/x86/kvm/x86.c struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt); kvm_vcpu 5478 arch/x86/kvm/x86.c struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt); kvm_vcpu 5485 arch/x86/kvm/x86.c struct kvm_vcpu *vcpu, u32 access, kvm_vcpu 5519 arch/x86/kvm/x86.c struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt); kvm_vcpu 5529 arch/x86/kvm/x86.c int kvm_write_guest_virt_system(struct kvm_vcpu *vcpu, gva_t addr, void *val, kvm_vcpu 5547 arch/x86/kvm/x86.c int handle_ud(struct kvm_vcpu *vcpu) kvm_vcpu 5565 arch/x86/kvm/x86.c static int vcpu_is_mmio_gpa(struct kvm_vcpu *vcpu, unsigned long gva, kvm_vcpu 5580 arch/x86/kvm/x86.c static int vcpu_mmio_gva_to_gpa(struct kvm_vcpu *vcpu, unsigned long gva, kvm_vcpu 5609 arch/x86/kvm/x86.c int emulator_write_phys(struct kvm_vcpu *vcpu, gpa_t gpa, kvm_vcpu 5622 arch/x86/kvm/x86.c int (*read_write_prepare)(struct kvm_vcpu *vcpu, void *val, kvm_vcpu 5624 arch/x86/kvm/x86.c int (*read_write_emulate)(struct kvm_vcpu *vcpu, gpa_t gpa, kvm_vcpu 5626 arch/x86/kvm/x86.c int (*read_write_mmio)(struct kvm_vcpu *vcpu, gpa_t gpa, kvm_vcpu 5628 arch/x86/kvm/x86.c int (*read_write_exit_mmio)(struct kvm_vcpu *vcpu, gpa_t gpa, kvm_vcpu 5633 arch/x86/kvm/x86.c static int read_prepare(struct kvm_vcpu *vcpu, void *val, int bytes) kvm_vcpu 5645 arch/x86/kvm/x86.c static int read_emulate(struct kvm_vcpu *vcpu, gpa_t gpa, kvm_vcpu 5651 arch/x86/kvm/x86.c static int write_emulate(struct kvm_vcpu *vcpu, gpa_t gpa, kvm_vcpu 5657 arch/x86/kvm/x86.c static int write_mmio(struct kvm_vcpu *vcpu, gpa_t gpa, int bytes, void *val) kvm_vcpu 5663 arch/x86/kvm/x86.c static int read_exit_mmio(struct kvm_vcpu *vcpu, gpa_t gpa, kvm_vcpu 5670 arch/x86/kvm/x86.c static int write_exit_mmio(struct kvm_vcpu *vcpu, gpa_t gpa, kvm_vcpu 5696 arch/x86/kvm/x86.c struct kvm_vcpu *vcpu, kvm_vcpu 5751 arch/x86/kvm/x86.c struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt); kvm_vcpu 5837 arch/x86/kvm/x86.c struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt); kvm_vcpu 5892 arch/x86/kvm/x86.c static int kernel_pio(struct kvm_vcpu *vcpu, void *pd) kvm_vcpu 5911 arch/x86/kvm/x86.c static int emulator_pio_in_out(struct kvm_vcpu *vcpu, int size, kvm_vcpu 5939 arch/x86/kvm/x86.c struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt); kvm_vcpu 5963 arch/x86/kvm/x86.c struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt); kvm_vcpu 5970 arch/x86/kvm/x86.c static unsigned long get_segment_base(struct kvm_vcpu *vcpu, int seg) kvm_vcpu 5980 arch/x86/kvm/x86.c static int kvm_emulate_wbinvd_noskip(struct kvm_vcpu *vcpu) kvm_vcpu 5998 arch/x86/kvm/x86.c int kvm_emulate_wbinvd(struct kvm_vcpu *vcpu) kvm_vcpu 6032 arch/x86/kvm/x86.c struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt); kvm_vcpu 6061 arch/x86/kvm/x86.c struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt); kvm_vcpu 6159 arch/x86/kvm/x86.c struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt); kvm_vcpu 6199 arch/x86/kvm/x86.c struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt); kvm_vcpu 6206 arch/x86/kvm/x86.c struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt); kvm_vcpu 6326 arch/x86/kvm/x86.c static void toggle_interruptibility(struct kvm_vcpu *vcpu, u32 mask) kvm_vcpu 6345 arch/x86/kvm/x86.c static bool inject_emulated_exception(struct kvm_vcpu *vcpu) kvm_vcpu 6359 arch/x86/kvm/x86.c static void init_emulate_ctxt(struct kvm_vcpu *vcpu) kvm_vcpu 6383 arch/x86/kvm/x86.c void kvm_inject_realmode_interrupt(struct kvm_vcpu *vcpu, int irq, int inc_eip) kvm_vcpu 6405 arch/x86/kvm/x86.c static int handle_emulation_failure(struct kvm_vcpu *vcpu, int emulation_type) kvm_vcpu 6434 arch/x86/kvm/x86.c static bool reexecute_instruction(struct kvm_vcpu *vcpu, gpa_t cr2_or_gpa, kvm_vcpu 6511 arch/x86/kvm/x86.c struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt); kvm_vcpu 6555 arch/x86/kvm/x86.c static int complete_emulated_mmio(struct kvm_vcpu *vcpu); kvm_vcpu 6556 arch/x86/kvm/x86.c static int complete_emulated_pio(struct kvm_vcpu *vcpu); kvm_vcpu 6558 arch/x86/kvm/x86.c static void kvm_smm_changed(struct kvm_vcpu *vcpu) kvm_vcpu 6586 arch/x86/kvm/x86.c static int kvm_vcpu_do_singlestep(struct kvm_vcpu *vcpu) kvm_vcpu 6601 arch/x86/kvm/x86.c int kvm_skip_emulated_instruction(struct kvm_vcpu *vcpu) kvm_vcpu 6624 arch/x86/kvm/x86.c static bool kvm_vcpu_check_breakpoint(struct kvm_vcpu *vcpu, int *r) kvm_vcpu 6694 arch/x86/kvm/x86.c int x86_emulate_instruction(struct kvm_vcpu *vcpu, gpa_t cr2_or_gpa, kvm_vcpu 6855 arch/x86/kvm/x86.c int kvm_emulate_instruction(struct kvm_vcpu *vcpu, int emulation_type) kvm_vcpu 6861 arch/x86/kvm/x86.c int kvm_emulate_instruction_from_buffer(struct kvm_vcpu *vcpu, kvm_vcpu 6868 arch/x86/kvm/x86.c static int complete_fast_pio_out_port_0x7e(struct kvm_vcpu *vcpu) kvm_vcpu 6874 arch/x86/kvm/x86.c static int complete_fast_pio_out(struct kvm_vcpu *vcpu) kvm_vcpu 6884 arch/x86/kvm/x86.c static int kvm_fast_pio_out(struct kvm_vcpu *vcpu, int size, kvm_vcpu 6909 arch/x86/kvm/x86.c static int complete_fast_pio_in(struct kvm_vcpu *vcpu) kvm_vcpu 6935 arch/x86/kvm/x86.c static int kvm_fast_pio_in(struct kvm_vcpu *vcpu, int size, kvm_vcpu 6957 arch/x86/kvm/x86.c int kvm_fast_pio(struct kvm_vcpu *vcpu, int size, unsigned short port, int in) kvm_vcpu 6993 arch/x86/kvm/x86.c struct kvm_vcpu *vcpu; kvm_vcpu 7029 arch/x86/kvm/x86.c struct kvm_vcpu *vcpu; kvm_vcpu 7153 arch/x86/kvm/x86.c DEFINE_PER_CPU(struct kvm_vcpu *, current_vcpu); kvm_vcpu 7183 arch/x86/kvm/x86.c struct kvm_vcpu *vcpu = __this_cpu_read(current_vcpu); kvm_vcpu 7202 arch/x86/kvm/x86.c struct kvm_vcpu *vcpu; kvm_vcpu 7347 arch/x86/kvm/x86.c int kvm_vcpu_halt(struct kvm_vcpu *vcpu) kvm_vcpu 7360 arch/x86/kvm/x86.c int kvm_emulate_halt(struct kvm_vcpu *vcpu) kvm_vcpu 7372 arch/x86/kvm/x86.c static int kvm_pv_clock_pairing(struct kvm_vcpu *vcpu, gpa_t paddr, kvm_vcpu 7420 arch/x86/kvm/x86.c void kvm_vcpu_deactivate_apicv(struct kvm_vcpu *vcpu) kvm_vcpu 7435 arch/x86/kvm/x86.c struct kvm_vcpu *target = NULL; kvm_vcpu 7450 arch/x86/kvm/x86.c int kvm_emulate_hypercall(struct kvm_vcpu *vcpu) kvm_vcpu 7517 arch/x86/kvm/x86.c struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt); kvm_vcpu 7527 arch/x86/kvm/x86.c static int dm_request_for_irq_injection(struct kvm_vcpu *vcpu) kvm_vcpu 7533 arch/x86/kvm/x86.c static void post_kvm_run_save(struct kvm_vcpu *vcpu) kvm_vcpu 7546 arch/x86/kvm/x86.c static void update_cr8_intercept(struct kvm_vcpu *vcpu) kvm_vcpu 7572 arch/x86/kvm/x86.c static int inject_pending_event(struct kvm_vcpu *vcpu) kvm_vcpu 7684 arch/x86/kvm/x86.c static void process_nmi(struct kvm_vcpu *vcpu) kvm_vcpu 7715 arch/x86/kvm/x86.c static void enter_smm_save_seg_32(struct kvm_vcpu *vcpu, char *buf, int n) kvm_vcpu 7734 arch/x86/kvm/x86.c static void enter_smm_save_seg_64(struct kvm_vcpu *vcpu, char *buf, int n) kvm_vcpu 7751 arch/x86/kvm/x86.c static void enter_smm_save_state_32(struct kvm_vcpu *vcpu, char *buf) kvm_vcpu 7802 arch/x86/kvm/x86.c static void enter_smm_save_state_64(struct kvm_vcpu *vcpu, char *buf) kvm_vcpu 7856 arch/x86/kvm/x86.c static void enter_smm(struct kvm_vcpu *vcpu) kvm_vcpu 7936 arch/x86/kvm/x86.c static void process_smi(struct kvm_vcpu *vcpu) kvm_vcpu 7947 arch/x86/kvm/x86.c static void vcpu_scan_ioapic(struct kvm_vcpu *vcpu) kvm_vcpu 7969 arch/x86/kvm/x86.c static void vcpu_load_eoi_exitmap(struct kvm_vcpu *vcpu) kvm_vcpu 7995 arch/x86/kvm/x86.c void kvm_vcpu_reload_apic_access_page(struct kvm_vcpu *vcpu) kvm_vcpu 8018 arch/x86/kvm/x86.c void __kvm_request_immediate_exit(struct kvm_vcpu *vcpu) kvm_vcpu 8029 arch/x86/kvm/x86.c static int vcpu_enter_guest(struct kvm_vcpu *vcpu) kvm_vcpu 8340 arch/x86/kvm/x86.c static inline int vcpu_block(struct kvm *kvm, struct kvm_vcpu *vcpu) kvm_vcpu 8374 arch/x86/kvm/x86.c static inline bool kvm_vcpu_running(struct kvm_vcpu *vcpu) kvm_vcpu 8383 arch/x86/kvm/x86.c static int vcpu_run(struct kvm_vcpu *vcpu) kvm_vcpu 8433 arch/x86/kvm/x86.c static inline int complete_emulated_io(struct kvm_vcpu *vcpu) kvm_vcpu 8443 arch/x86/kvm/x86.c static int complete_emulated_pio(struct kvm_vcpu *vcpu) kvm_vcpu 8468 arch/x86/kvm/x86.c static int complete_emulated_mmio(struct kvm_vcpu *vcpu) kvm_vcpu 8527 arch/x86/kvm/x86.c static void kvm_load_guest_fpu(struct kvm_vcpu *vcpu) kvm_vcpu 8544 arch/x86/kvm/x86.c static void kvm_put_guest_fpu(struct kvm_vcpu *vcpu) kvm_vcpu 8559 arch/x86/kvm/x86.c int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run) kvm_vcpu 8604 arch/x86/kvm/x86.c int (*cui)(struct kvm_vcpu *) = vcpu->arch.complete_userspace_io; kvm_vcpu 8628 arch/x86/kvm/x86.c static void __get_regs(struct kvm_vcpu *vcpu, struct kvm_regs *regs) kvm_vcpu 8664 arch/x86/kvm/x86.c int kvm_arch_vcpu_ioctl_get_regs(struct kvm_vcpu *vcpu, struct kvm_regs *regs) kvm_vcpu 8672 arch/x86/kvm/x86.c static void __set_regs(struct kvm_vcpu *vcpu, struct kvm_regs *regs) kvm_vcpu 8704 arch/x86/kvm/x86.c int kvm_arch_vcpu_ioctl_set_regs(struct kvm_vcpu *vcpu, struct kvm_regs *regs) kvm_vcpu 8712 arch/x86/kvm/x86.c void kvm_get_cs_db_l_bits(struct kvm_vcpu *vcpu, int *db, int *l) kvm_vcpu 8722 arch/x86/kvm/x86.c static void __get_sregs(struct kvm_vcpu *vcpu, struct kvm_sregs *sregs) kvm_vcpu 8758 arch/x86/kvm/x86.c int kvm_arch_vcpu_ioctl_get_sregs(struct kvm_vcpu *vcpu, kvm_vcpu 8767 arch/x86/kvm/x86.c int kvm_arch_vcpu_ioctl_get_mpstate(struct kvm_vcpu *vcpu, kvm_vcpu 8787 arch/x86/kvm/x86.c int kvm_arch_vcpu_ioctl_set_mpstate(struct kvm_vcpu *vcpu, kvm_vcpu 8817 arch/x86/kvm/x86.c int kvm_task_switch(struct kvm_vcpu *vcpu, u16 tss_selector, int idt_index, kvm_vcpu 8841 arch/x86/kvm/x86.c static int kvm_valid_sregs(struct kvm_vcpu *vcpu, struct kvm_sregs *sregs) kvm_vcpu 8864 arch/x86/kvm/x86.c static int __set_sregs(struct kvm_vcpu *vcpu, struct kvm_sregs *sregs) kvm_vcpu 8952 arch/x86/kvm/x86.c int kvm_arch_vcpu_ioctl_set_sregs(struct kvm_vcpu *vcpu, kvm_vcpu 8963 arch/x86/kvm/x86.c int kvm_arch_vcpu_ioctl_set_guest_debug(struct kvm_vcpu *vcpu, kvm_vcpu 9023 arch/x86/kvm/x86.c int kvm_arch_vcpu_ioctl_translate(struct kvm_vcpu *vcpu, kvm_vcpu 9044 arch/x86/kvm/x86.c int kvm_arch_vcpu_ioctl_get_fpu(struct kvm_vcpu *vcpu, struct kvm_fpu *fpu) kvm_vcpu 9064 arch/x86/kvm/x86.c int kvm_arch_vcpu_ioctl_set_fpu(struct kvm_vcpu *vcpu, struct kvm_fpu *fpu) kvm_vcpu 9085 arch/x86/kvm/x86.c static void store_regs(struct kvm_vcpu *vcpu) kvm_vcpu 9100 arch/x86/kvm/x86.c static int sync_regs(struct kvm_vcpu *vcpu) kvm_vcpu 9124 arch/x86/kvm/x86.c static void fx_init(struct kvm_vcpu *vcpu) kvm_vcpu 9139 arch/x86/kvm/x86.c void kvm_arch_vcpu_free(struct kvm_vcpu *vcpu) kvm_vcpu 9152 arch/x86/kvm/x86.c struct kvm_vcpu *kvm_arch_vcpu_create(struct kvm *kvm, kvm_vcpu 9155 arch/x86/kvm/x86.c struct kvm_vcpu *vcpu; kvm_vcpu 9167 arch/x86/kvm/x86.c int kvm_arch_vcpu_setup(struct kvm_vcpu *vcpu) kvm_vcpu 9179 arch/x86/kvm/x86.c void kvm_arch_vcpu_postcreate(struct kvm_vcpu *vcpu) kvm_vcpu 9207 arch/x86/kvm/x86.c void kvm_arch_vcpu_destroy(struct kvm_vcpu *vcpu) kvm_vcpu 9212 arch/x86/kvm/x86.c void kvm_vcpu_reset(struct kvm_vcpu *vcpu, bool init_event) kvm_vcpu 9285 arch/x86/kvm/x86.c void kvm_vcpu_deliver_sipi_vector(struct kvm_vcpu *vcpu, u8 vector) kvm_vcpu 9299 arch/x86/kvm/x86.c struct kvm_vcpu *vcpu; kvm_vcpu 9431 arch/x86/kvm/x86.c bool kvm_vcpu_is_reset_bsp(struct kvm_vcpu *vcpu) kvm_vcpu 9437 arch/x86/kvm/x86.c bool kvm_vcpu_is_bsp(struct kvm_vcpu *vcpu) kvm_vcpu 9445 arch/x86/kvm/x86.c int kvm_arch_vcpu_init(struct kvm_vcpu *vcpu) kvm_vcpu 9521 arch/x86/kvm/x86.c void kvm_arch_vcpu_uninit(struct kvm_vcpu *vcpu) kvm_vcpu 9537 arch/x86/kvm/x86.c void kvm_arch_sched_in(struct kvm_vcpu *vcpu, int cpu) kvm_vcpu 9585 arch/x86/kvm/x86.c static void kvm_unload_vcpu_mmu(struct kvm_vcpu *vcpu) kvm_vcpu 9595 arch/x86/kvm/x86.c struct kvm_vcpu *vcpu; kvm_vcpu 9812 arch/x86/kvm/x86.c struct kvm_vcpu *vcpu; kvm_vcpu 9945 arch/x86/kvm/x86.c static inline bool kvm_guest_apic_has_interrupt(struct kvm_vcpu *vcpu) kvm_vcpu 9952 arch/x86/kvm/x86.c static inline bool kvm_vcpu_has_events(struct kvm_vcpu *vcpu) kvm_vcpu 9986 arch/x86/kvm/x86.c int kvm_arch_vcpu_runnable(struct kvm_vcpu *vcpu) kvm_vcpu 9991 arch/x86/kvm/x86.c bool kvm_arch_dy_runnable(struct kvm_vcpu *vcpu) kvm_vcpu 10007 arch/x86/kvm/x86.c bool kvm_arch_vcpu_in_kernel(struct kvm_vcpu *vcpu) kvm_vcpu 10012 arch/x86/kvm/x86.c int kvm_arch_vcpu_should_kick(struct kvm_vcpu *vcpu) kvm_vcpu 10017 arch/x86/kvm/x86.c int kvm_arch_interrupt_allowed(struct kvm_vcpu *vcpu) kvm_vcpu 10022 arch/x86/kvm/x86.c unsigned long kvm_get_linear_rip(struct kvm_vcpu *vcpu) kvm_vcpu 10031 arch/x86/kvm/x86.c bool kvm_is_linear_rip(struct kvm_vcpu *vcpu, unsigned long linear_rip) kvm_vcpu 10037 arch/x86/kvm/x86.c unsigned long kvm_get_rflags(struct kvm_vcpu *vcpu) kvm_vcpu 10048 arch/x86/kvm/x86.c static void __kvm_set_rflags(struct kvm_vcpu *vcpu, unsigned long rflags) kvm_vcpu 10056 arch/x86/kvm/x86.c void kvm_set_rflags(struct kvm_vcpu *vcpu, unsigned long rflags) kvm_vcpu 10063 arch/x86/kvm/x86.c void kvm_arch_async_page_ready(struct kvm_vcpu *vcpu, struct kvm_async_pf *work) kvm_vcpu 10092 arch/x86/kvm/x86.c static void kvm_add_async_pf_gfn(struct kvm_vcpu *vcpu, gfn_t gfn) kvm_vcpu 10102 arch/x86/kvm/x86.c static u32 kvm_async_pf_gfn_slot(struct kvm_vcpu *vcpu, gfn_t gfn) kvm_vcpu 10115 arch/x86/kvm/x86.c bool kvm_find_async_pf_gfn(struct kvm_vcpu *vcpu, gfn_t gfn) kvm_vcpu 10120 arch/x86/kvm/x86.c static void kvm_del_async_pf_gfn(struct kvm_vcpu *vcpu, gfn_t gfn) kvm_vcpu 10143 arch/x86/kvm/x86.c static int apf_put_user(struct kvm_vcpu *vcpu, u32 val) kvm_vcpu 10150 arch/x86/kvm/x86.c static int apf_get_user(struct kvm_vcpu *vcpu, u32 *val) kvm_vcpu 10157 arch/x86/kvm/x86.c static bool kvm_can_deliver_async_pf(struct kvm_vcpu *vcpu) kvm_vcpu 10170 arch/x86/kvm/x86.c bool kvm_can_do_async_pf(struct kvm_vcpu *vcpu) kvm_vcpu 10187 arch/x86/kvm/x86.c void kvm_arch_async_page_not_present(struct kvm_vcpu *vcpu, kvm_vcpu 10217 arch/x86/kvm/x86.c void kvm_arch_async_page_present(struct kvm_vcpu *vcpu, kvm_vcpu 10256 arch/x86/kvm/x86.c bool kvm_arch_can_inject_async_page_present(struct kvm_vcpu *vcpu) kvm_vcpu 10351 arch/x86/kvm/x86.c bool kvm_arch_no_poll(struct kvm_vcpu *vcpu) kvm_vcpu 49 arch/x86/kvm/x86.h static inline void kvm_clear_exception_queue(struct kvm_vcpu *vcpu) kvm_vcpu 55 arch/x86/kvm/x86.h static inline void kvm_queue_interrupt(struct kvm_vcpu *vcpu, u8 vector, kvm_vcpu 63 arch/x86/kvm/x86.h static inline void kvm_clear_interrupt_queue(struct kvm_vcpu *vcpu) kvm_vcpu 68 arch/x86/kvm/x86.h static inline bool kvm_event_needs_reinjection(struct kvm_vcpu *vcpu) kvm_vcpu 79 arch/x86/kvm/x86.h static inline bool is_protmode(struct kvm_vcpu *vcpu) kvm_vcpu 84 arch/x86/kvm/x86.h static inline int is_long_mode(struct kvm_vcpu *vcpu) kvm_vcpu 93 arch/x86/kvm/x86.h static inline bool is_64_bit_mode(struct kvm_vcpu *vcpu) kvm_vcpu 103 arch/x86/kvm/x86.h static inline bool is_la57_mode(struct kvm_vcpu *vcpu) kvm_vcpu 122 arch/x86/kvm/x86.h static inline bool mmu_is_nested(struct kvm_vcpu *vcpu) kvm_vcpu 127 arch/x86/kvm/x86.h static inline int is_pae(struct kvm_vcpu *vcpu) kvm_vcpu 132 arch/x86/kvm/x86.h static inline int is_pse(struct kvm_vcpu *vcpu) kvm_vcpu 137 arch/x86/kvm/x86.h static inline int is_paging(struct kvm_vcpu *vcpu) kvm_vcpu 142 arch/x86/kvm/x86.h static inline bool is_pae_paging(struct kvm_vcpu *vcpu) kvm_vcpu 152 arch/x86/kvm/x86.h static inline u8 vcpu_virt_addr_bits(struct kvm_vcpu *vcpu) kvm_vcpu 167 arch/x86/kvm/x86.h static inline bool is_noncanonical_address(u64 la, struct kvm_vcpu *vcpu) kvm_vcpu 186 arch/x86/kvm/x86.h static inline void vcpu_cache_mmio_info(struct kvm_vcpu *vcpu, kvm_vcpu 204 arch/x86/kvm/x86.h static inline bool vcpu_match_mmio_gen(struct kvm_vcpu *vcpu) kvm_vcpu 215 arch/x86/kvm/x86.h static inline void vcpu_clear_mmio_info(struct kvm_vcpu *vcpu, gva_t gva) kvm_vcpu 223 arch/x86/kvm/x86.h static inline bool vcpu_match_mmio_gva(struct kvm_vcpu *vcpu, unsigned long gva) kvm_vcpu 232 arch/x86/kvm/x86.h static inline bool vcpu_match_mmio_gpa(struct kvm_vcpu *vcpu, gpa_t gpa) kvm_vcpu 241 arch/x86/kvm/x86.h static inline unsigned long kvm_register_readl(struct kvm_vcpu *vcpu, kvm_vcpu 249 arch/x86/kvm/x86.h static inline void kvm_register_writel(struct kvm_vcpu *vcpu, kvm_vcpu 263 arch/x86/kvm/x86.h void kvm_set_pending_timer(struct kvm_vcpu *vcpu); kvm_vcpu 264 arch/x86/kvm/x86.h void kvm_inject_realmode_interrupt(struct kvm_vcpu *vcpu, int irq, int inc_eip); kvm_vcpu 266 arch/x86/kvm/x86.h void kvm_write_tsc(struct kvm_vcpu *vcpu, struct msr_data *msr); kvm_vcpu 269 arch/x86/kvm/x86.h int kvm_read_guest_virt(struct kvm_vcpu *vcpu, kvm_vcpu 273 arch/x86/kvm/x86.h int kvm_write_guest_virt_system(struct kvm_vcpu *vcpu, kvm_vcpu 277 arch/x86/kvm/x86.h int handle_ud(struct kvm_vcpu *vcpu); kvm_vcpu 279 arch/x86/kvm/x86.h void kvm_deliver_exception_payload(struct kvm_vcpu *vcpu); kvm_vcpu 281 arch/x86/kvm/x86.h void kvm_vcpu_mtrr_init(struct kvm_vcpu *vcpu); kvm_vcpu 282 arch/x86/kvm/x86.h u8 kvm_mtrr_get_guest_memory_type(struct kvm_vcpu *vcpu, gfn_t gfn); kvm_vcpu 283 arch/x86/kvm/x86.h bool kvm_mtrr_valid(struct kvm_vcpu *vcpu, u32 msr, u64 data); kvm_vcpu 284 arch/x86/kvm/x86.h int kvm_mtrr_set_msr(struct kvm_vcpu *vcpu, u32 msr, u64 data); kvm_vcpu 285 arch/x86/kvm/x86.h int kvm_mtrr_get_msr(struct kvm_vcpu *vcpu, u32 msr, u64 *pdata); kvm_vcpu 286 arch/x86/kvm/x86.h bool kvm_mtrr_check_gfn_range_consistency(struct kvm_vcpu *vcpu, gfn_t gfn, kvm_vcpu 289 arch/x86/kvm/x86.h int x86_emulate_instruction(struct kvm_vcpu *vcpu, gpa_t cr2_or_gpa, kvm_vcpu 308 arch/x86/kvm/x86.h static inline u64 nsec_to_cycles(struct kvm_vcpu *vcpu, u64 nsec) kvm_vcpu 348 arch/x86/kvm/x86.h DECLARE_PER_CPU(struct kvm_vcpu *, current_vcpu); kvm_vcpu 350 arch/x86/kvm/x86.h static inline void kvm_before_interrupt(struct kvm_vcpu *vcpu) kvm_vcpu 355 arch/x86/kvm/x86.h static inline void kvm_after_interrupt(struct kvm_vcpu *vcpu) kvm_vcpu 369 arch/x86/kvm/x86.h void kvm_load_guest_xcr0(struct kvm_vcpu *vcpu); kvm_vcpu 370 arch/x86/kvm/x86.h void kvm_put_guest_xcr0(struct kvm_vcpu *vcpu); kvm_vcpu 1702 drivers/gpu/drm/i915/gvt/kvmgt.c static void kvmgt_page_track_write(struct kvm_vcpu *vcpu, gpa_t gpa, kvm_vcpu 275 drivers/s390/crypto/vfio_ap_ops.c static int handle_pqap(struct kvm_vcpu *vcpu) kvm_vcpu 27 include/kvm/arm_arch_timer.h struct kvm_vcpu *vcpu; kvm_vcpu 71 include/kvm/arm_arch_timer.h int kvm_timer_enable(struct kvm_vcpu *vcpu); kvm_vcpu 72 include/kvm/arm_arch_timer.h int kvm_timer_vcpu_reset(struct kvm_vcpu *vcpu); kvm_vcpu 73 include/kvm/arm_arch_timer.h void kvm_timer_vcpu_init(struct kvm_vcpu *vcpu); kvm_vcpu 74 include/kvm/arm_arch_timer.h void kvm_timer_sync_hwstate(struct kvm_vcpu *vcpu); kvm_vcpu 75 include/kvm/arm_arch_timer.h bool kvm_timer_should_notify_user(struct kvm_vcpu *vcpu); kvm_vcpu 76 include/kvm/arm_arch_timer.h void kvm_timer_update_run(struct kvm_vcpu *vcpu); kvm_vcpu 77 include/kvm/arm_arch_timer.h void kvm_timer_vcpu_terminate(struct kvm_vcpu *vcpu); kvm_vcpu 79 include/kvm/arm_arch_timer.h u64 kvm_arm_timer_get_reg(struct kvm_vcpu *, u64 regid); kvm_vcpu 80 include/kvm/arm_arch_timer.h int kvm_arm_timer_set_reg(struct kvm_vcpu *, u64 regid, u64 value); kvm_vcpu 82 include/kvm/arm_arch_timer.h int kvm_arm_timer_set_attr(struct kvm_vcpu *vcpu, struct kvm_device_attr *attr); kvm_vcpu 83 include/kvm/arm_arch_timer.h int kvm_arm_timer_get_attr(struct kvm_vcpu *vcpu, struct kvm_device_attr *attr); kvm_vcpu 84 include/kvm/arm_arch_timer.h int kvm_arm_timer_has_attr(struct kvm_vcpu *vcpu, struct kvm_device_attr *attr); kvm_vcpu 86 include/kvm/arm_arch_timer.h bool kvm_timer_is_pending(struct kvm_vcpu *vcpu); kvm_vcpu 90 include/kvm/arm_arch_timer.h void kvm_timer_vcpu_load(struct kvm_vcpu *vcpu); kvm_vcpu 91 include/kvm/arm_arch_timer.h void kvm_timer_vcpu_put(struct kvm_vcpu *vcpu); kvm_vcpu 104 include/kvm/arm_arch_timer.h u64 kvm_arm_timer_read_sysreg(struct kvm_vcpu *vcpu, kvm_vcpu 107 include/kvm/arm_arch_timer.h void kvm_arm_timer_write_sysreg(struct kvm_vcpu *vcpu, kvm_vcpu 34 include/kvm/arm_pmu.h u64 kvm_pmu_get_counter_value(struct kvm_vcpu *vcpu, u64 select_idx); kvm_vcpu 35 include/kvm/arm_pmu.h void kvm_pmu_set_counter_value(struct kvm_vcpu *vcpu, u64 select_idx, u64 val); kvm_vcpu 36 include/kvm/arm_pmu.h u64 kvm_pmu_valid_counter_mask(struct kvm_vcpu *vcpu); kvm_vcpu 37 include/kvm/arm_pmu.h void kvm_pmu_vcpu_init(struct kvm_vcpu *vcpu); kvm_vcpu 38 include/kvm/arm_pmu.h void kvm_pmu_vcpu_reset(struct kvm_vcpu *vcpu); kvm_vcpu 39 include/kvm/arm_pmu.h void kvm_pmu_vcpu_destroy(struct kvm_vcpu *vcpu); kvm_vcpu 40 include/kvm/arm_pmu.h void kvm_pmu_disable_counter_mask(struct kvm_vcpu *vcpu, u64 val); kvm_vcpu 41 include/kvm/arm_pmu.h void kvm_pmu_enable_counter_mask(struct kvm_vcpu *vcpu, u64 val); kvm_vcpu 42 include/kvm/arm_pmu.h void kvm_pmu_flush_hwstate(struct kvm_vcpu *vcpu); kvm_vcpu 43 include/kvm/arm_pmu.h void kvm_pmu_sync_hwstate(struct kvm_vcpu *vcpu); kvm_vcpu 44 include/kvm/arm_pmu.h bool kvm_pmu_should_notify_user(struct kvm_vcpu *vcpu); kvm_vcpu 45 include/kvm/arm_pmu.h void kvm_pmu_update_run(struct kvm_vcpu *vcpu); kvm_vcpu 46 include/kvm/arm_pmu.h void kvm_pmu_software_increment(struct kvm_vcpu *vcpu, u64 val); kvm_vcpu 47 include/kvm/arm_pmu.h void kvm_pmu_handle_pmcr(struct kvm_vcpu *vcpu, u64 val); kvm_vcpu 48 include/kvm/arm_pmu.h void kvm_pmu_set_counter_event_type(struct kvm_vcpu *vcpu, u64 data, kvm_vcpu 51 include/kvm/arm_pmu.h int kvm_arm_pmu_v3_set_attr(struct kvm_vcpu *vcpu, kvm_vcpu 53 include/kvm/arm_pmu.h int kvm_arm_pmu_v3_get_attr(struct kvm_vcpu *vcpu, kvm_vcpu 55 include/kvm/arm_pmu.h int kvm_arm_pmu_v3_has_attr(struct kvm_vcpu *vcpu, kvm_vcpu 57 include/kvm/arm_pmu.h int kvm_arm_pmu_v3_enable(struct kvm_vcpu *vcpu); kvm_vcpu 64 include/kvm/arm_pmu.h static inline u64 kvm_pmu_get_counter_value(struct kvm_vcpu *vcpu, kvm_vcpu 69 include/kvm/arm_pmu.h static inline void kvm_pmu_set_counter_value(struct kvm_vcpu *vcpu, kvm_vcpu 71 include/kvm/arm_pmu.h static inline u64 kvm_pmu_valid_counter_mask(struct kvm_vcpu *vcpu) kvm_vcpu 75 include/kvm/arm_pmu.h static inline void kvm_pmu_vcpu_init(struct kvm_vcpu *vcpu) {} kvm_vcpu 76 include/kvm/arm_pmu.h static inline void kvm_pmu_vcpu_reset(struct kvm_vcpu *vcpu) {} kvm_vcpu 77 include/kvm/arm_pmu.h static inline void kvm_pmu_vcpu_destroy(struct kvm_vcpu *vcpu) {} kvm_vcpu 78 include/kvm/arm_pmu.h static inline void kvm_pmu_disable_counter_mask(struct kvm_vcpu *vcpu, u64 val) {} kvm_vcpu 79 include/kvm/arm_pmu.h static inline void kvm_pmu_enable_counter_mask(struct kvm_vcpu *vcpu, u64 val) {} kvm_vcpu 80 include/kvm/arm_pmu.h static inline void kvm_pmu_flush_hwstate(struct kvm_vcpu *vcpu) {} kvm_vcpu 81 include/kvm/arm_pmu.h static inline void kvm_pmu_sync_hwstate(struct kvm_vcpu *vcpu) {} kvm_vcpu 82 include/kvm/arm_pmu.h static inline bool kvm_pmu_should_notify_user(struct kvm_vcpu *vcpu) kvm_vcpu 86 include/kvm/arm_pmu.h static inline void kvm_pmu_update_run(struct kvm_vcpu *vcpu) {} kvm_vcpu 87 include/kvm/arm_pmu.h static inline void kvm_pmu_software_increment(struct kvm_vcpu *vcpu, u64 val) {} kvm_vcpu 88 include/kvm/arm_pmu.h static inline void kvm_pmu_handle_pmcr(struct kvm_vcpu *vcpu, u64 val) {} kvm_vcpu 89 include/kvm/arm_pmu.h static inline void kvm_pmu_set_counter_event_type(struct kvm_vcpu *vcpu, kvm_vcpu 92 include/kvm/arm_pmu.h static inline int kvm_arm_pmu_v3_set_attr(struct kvm_vcpu *vcpu, kvm_vcpu 97 include/kvm/arm_pmu.h static inline int kvm_arm_pmu_v3_get_attr(struct kvm_vcpu *vcpu, kvm_vcpu 102 include/kvm/arm_pmu.h static inline int kvm_arm_pmu_v3_has_attr(struct kvm_vcpu *vcpu, kvm_vcpu 107 include/kvm/arm_pmu.h static inline int kvm_arm_pmu_v3_enable(struct kvm_vcpu *vcpu) kvm_vcpu 23 include/kvm/arm_psci.h static inline int kvm_psci_version(struct kvm_vcpu *vcpu, struct kvm *kvm) kvm_vcpu 43 include/kvm/arm_psci.h int kvm_hvc_call_handler(struct kvm_vcpu *vcpu); kvm_vcpu 47 include/kvm/arm_psci.h int kvm_arm_get_fw_num_regs(struct kvm_vcpu *vcpu); kvm_vcpu 48 include/kvm/arm_psci.h int kvm_arm_copy_fw_reg_indices(struct kvm_vcpu *vcpu, u64 __user *uindices); kvm_vcpu 49 include/kvm/arm_psci.h int kvm_arm_get_fw_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg); kvm_vcpu 50 include/kvm/arm_psci.h int kvm_arm_set_fw_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg); kvm_vcpu 96 include/kvm/arm_vgic.h struct kvm_vcpu *vcpu; /* SGIs and PPIs: The VCPU kvm_vcpu 101 include/kvm/arm_vgic.h struct kvm_vcpu *target_vcpu; /* The VCPU that this interrupt should kvm_vcpu 156 include/kvm/arm_vgic.h struct kvm_vcpu *redist_vcpu; kvm_vcpu 336 include/kvm/arm_vgic.h int kvm_vgic_vcpu_init(struct kvm_vcpu *vcpu); kvm_vcpu 339 include/kvm/arm_vgic.h void kvm_vgic_vcpu_destroy(struct kvm_vcpu *vcpu); kvm_vcpu 346 include/kvm/arm_vgic.h int kvm_vgic_map_phys_irq(struct kvm_vcpu *vcpu, unsigned int host_irq, kvm_vcpu 348 include/kvm/arm_vgic.h int kvm_vgic_unmap_phys_irq(struct kvm_vcpu *vcpu, unsigned int vintid); kvm_vcpu 349 include/kvm/arm_vgic.h bool kvm_vgic_map_is_active(struct kvm_vcpu *vcpu, unsigned int vintid); kvm_vcpu 351 include/kvm/arm_vgic.h int kvm_vgic_vcpu_pending_irq(struct kvm_vcpu *vcpu); kvm_vcpu 353 include/kvm/arm_vgic.h void kvm_vgic_load(struct kvm_vcpu *vcpu); kvm_vcpu 354 include/kvm/arm_vgic.h void kvm_vgic_put(struct kvm_vcpu *vcpu); kvm_vcpu 355 include/kvm/arm_vgic.h void kvm_vgic_vmcr_sync(struct kvm_vcpu *vcpu); kvm_vcpu 363 include/kvm/arm_vgic.h bool kvm_vcpu_has_pending_irqs(struct kvm_vcpu *vcpu); kvm_vcpu 364 include/kvm/arm_vgic.h void kvm_vgic_sync_hwstate(struct kvm_vcpu *vcpu); kvm_vcpu 365 include/kvm/arm_vgic.h void kvm_vgic_flush_hwstate(struct kvm_vcpu *vcpu); kvm_vcpu 366 include/kvm/arm_vgic.h void kvm_vgic_reset_mapped_irq(struct kvm_vcpu *vcpu, u32 vintid); kvm_vcpu 368 include/kvm/arm_vgic.h void vgic_v3_dispatch_sgi(struct kvm_vcpu *vcpu, u64 reg, bool allow_group1); kvm_vcpu 389 include/kvm/arm_vgic.h int kvm_vgic_set_owner(struct kvm_vcpu *vcpu, unsigned int intid, void *owner); kvm_vcpu 399 include/kvm/arm_vgic.h void kvm_vgic_v4_enable_doorbell(struct kvm_vcpu *vcpu); kvm_vcpu 400 include/kvm/arm_vgic.h void kvm_vgic_v4_disable_doorbell(struct kvm_vcpu *vcpu); kvm_vcpu 10 include/kvm/iodev.h struct kvm_vcpu; kvm_vcpu 18 include/kvm/iodev.h int (*read)(struct kvm_vcpu *vcpu, kvm_vcpu 23 include/kvm/iodev.h int (*write)(struct kvm_vcpu *vcpu, kvm_vcpu 42 include/kvm/iodev.h static inline int kvm_iodevice_read(struct kvm_vcpu *vcpu, kvm_vcpu 50 include/kvm/iodev.h static inline int kvm_iodevice_write(struct kvm_vcpu *vcpu, kvm_vcpu 152 include/linux/kvm_host.h BUILD_BUG_ON((unsigned)(nr) >= (FIELD_SIZEOF(struct kvm_vcpu, requests) * 8) - KVM_REQUEST_ARCH_BASE); \ kvm_vcpu 187 include/linux/kvm_host.h int kvm_io_bus_write(struct kvm_vcpu *vcpu, enum kvm_bus bus_idx, gpa_t addr, kvm_vcpu 189 include/linux/kvm_host.h int kvm_io_bus_write_cookie(struct kvm_vcpu *vcpu, enum kvm_bus bus_idx, kvm_vcpu 191 include/linux/kvm_host.h int kvm_io_bus_read(struct kvm_vcpu *vcpu, enum kvm_bus bus_idx, gpa_t addr, kvm_vcpu 205 include/linux/kvm_host.h struct kvm_vcpu *vcpu; kvm_vcpu 213 include/linux/kvm_host.h void kvm_clear_async_pf_completion_queue(struct kvm_vcpu *vcpu); kvm_vcpu 214 include/linux/kvm_host.h void kvm_check_async_pf_completion(struct kvm_vcpu *vcpu); kvm_vcpu 215 include/linux/kvm_host.h int kvm_setup_async_pf(struct kvm_vcpu *vcpu, gpa_t cr2_or_gpa, kvm_vcpu 217 include/linux/kvm_host.h int kvm_async_pf_wakeup_all(struct kvm_vcpu *vcpu); kvm_vcpu 326 include/linux/kvm_host.h static inline int kvm_vcpu_exiting_guest_mode(struct kvm_vcpu *vcpu) kvm_vcpu 423 include/linux/kvm_host.h static inline int kvm_arch_vcpu_memslots_id(struct kvm_vcpu *vcpu) kvm_vcpu 448 include/linux/kvm_host.h struct kvm_vcpu *vcpus[KVM_MAX_VCPUS]; kvm_vcpu 539 include/linux/kvm_host.h static inline struct kvm_vcpu *kvm_get_vcpu(struct kvm *kvm, int i) kvm_vcpu 555 include/linux/kvm_host.h static inline struct kvm_vcpu *kvm_get_vcpu_by_id(struct kvm *kvm, int id) kvm_vcpu 557 include/linux/kvm_host.h struct kvm_vcpu *vcpu = NULL; kvm_vcpu 572 include/linux/kvm_host.h static inline int kvm_vcpu_get_idx(struct kvm_vcpu *vcpu) kvm_vcpu 574 include/linux/kvm_host.h struct kvm_vcpu *tmp; kvm_vcpu 588 include/linux/kvm_host.h int kvm_vcpu_init(struct kvm_vcpu *vcpu, struct kvm *kvm, unsigned id); kvm_vcpu 589 include/linux/kvm_host.h void kvm_vcpu_uninit(struct kvm_vcpu *vcpu); kvm_vcpu 591 include/linux/kvm_host.h void vcpu_load(struct kvm_vcpu *vcpu); kvm_vcpu 592 include/linux/kvm_host.h void vcpu_put(struct kvm_vcpu *vcpu); kvm_vcpu 639 include/linux/kvm_host.h static inline struct kvm_memslots *kvm_vcpu_memslots(struct kvm_vcpu *vcpu) kvm_vcpu 754 include/linux/kvm_host.h unsigned long kvm_host_page_size(struct kvm_vcpu *vcpu, gfn_t gfn); kvm_vcpu 757 include/linux/kvm_host.h struct kvm_memslots *kvm_vcpu_memslots(struct kvm_vcpu *vcpu); kvm_vcpu 758 include/linux/kvm_host.h struct kvm_memory_slot *kvm_vcpu_gfn_to_memslot(struct kvm_vcpu *vcpu, gfn_t gfn); kvm_vcpu 759 include/linux/kvm_host.h kvm_pfn_t kvm_vcpu_gfn_to_pfn_atomic(struct kvm_vcpu *vcpu, gfn_t gfn); kvm_vcpu 760 include/linux/kvm_host.h kvm_pfn_t kvm_vcpu_gfn_to_pfn(struct kvm_vcpu *vcpu, gfn_t gfn); kvm_vcpu 761 include/linux/kvm_host.h int kvm_vcpu_map(struct kvm_vcpu *vcpu, gpa_t gpa, struct kvm_host_map *map); kvm_vcpu 762 include/linux/kvm_host.h int kvm_map_gfn(struct kvm_vcpu *vcpu, gfn_t gfn, struct kvm_host_map *map, kvm_vcpu 764 include/linux/kvm_host.h struct page *kvm_vcpu_gfn_to_page(struct kvm_vcpu *vcpu, gfn_t gfn); kvm_vcpu 765 include/linux/kvm_host.h void kvm_vcpu_unmap(struct kvm_vcpu *vcpu, struct kvm_host_map *map, bool dirty); kvm_vcpu 766 include/linux/kvm_host.h int kvm_unmap_gfn(struct kvm_vcpu *vcpu, struct kvm_host_map *map, kvm_vcpu 768 include/linux/kvm_host.h unsigned long kvm_vcpu_gfn_to_hva(struct kvm_vcpu *vcpu, gfn_t gfn); kvm_vcpu 769 include/linux/kvm_host.h unsigned long kvm_vcpu_gfn_to_hva_prot(struct kvm_vcpu *vcpu, gfn_t gfn, bool *writable); kvm_vcpu 770 include/linux/kvm_host.h int kvm_vcpu_read_guest_page(struct kvm_vcpu *vcpu, gfn_t gfn, void *data, int offset, kvm_vcpu 772 include/linux/kvm_host.h int kvm_vcpu_read_guest_atomic(struct kvm_vcpu *vcpu, gpa_t gpa, void *data, kvm_vcpu 774 include/linux/kvm_host.h int kvm_vcpu_read_guest(struct kvm_vcpu *vcpu, gpa_t gpa, void *data, kvm_vcpu 776 include/linux/kvm_host.h int kvm_vcpu_write_guest_page(struct kvm_vcpu *vcpu, gfn_t gfn, const void *data, kvm_vcpu 778 include/linux/kvm_host.h int kvm_vcpu_write_guest(struct kvm_vcpu *vcpu, gpa_t gpa, const void *data, kvm_vcpu 780 include/linux/kvm_host.h void kvm_vcpu_mark_page_dirty(struct kvm_vcpu *vcpu, gfn_t gfn); kvm_vcpu 782 include/linux/kvm_host.h void kvm_sigset_activate(struct kvm_vcpu *vcpu); kvm_vcpu 783 include/linux/kvm_host.h void kvm_sigset_deactivate(struct kvm_vcpu *vcpu); kvm_vcpu 785 include/linux/kvm_host.h void kvm_vcpu_block(struct kvm_vcpu *vcpu); kvm_vcpu 786 include/linux/kvm_host.h void kvm_arch_vcpu_blocking(struct kvm_vcpu *vcpu); kvm_vcpu 787 include/linux/kvm_host.h void kvm_arch_vcpu_unblocking(struct kvm_vcpu *vcpu); kvm_vcpu 788 include/linux/kvm_host.h bool kvm_vcpu_wake_up(struct kvm_vcpu *vcpu); kvm_vcpu 789 include/linux/kvm_host.h void kvm_vcpu_kick(struct kvm_vcpu *vcpu); kvm_vcpu 790 include/linux/kvm_host.h int kvm_vcpu_yield_to(struct kvm_vcpu *target); kvm_vcpu 791 include/linux/kvm_host.h void kvm_vcpu_on_spin(struct kvm_vcpu *vcpu, bool usermode_vcpu_not_eligible); kvm_vcpu 804 include/linux/kvm_host.h vm_fault_t kvm_arch_vcpu_fault(struct kvm_vcpu *vcpu, struct vm_fault *vmf); kvm_vcpu 833 include/linux/kvm_host.h int kvm_arch_vcpu_ioctl_get_fpu(struct kvm_vcpu *vcpu, struct kvm_fpu *fpu); kvm_vcpu 834 include/linux/kvm_host.h int kvm_arch_vcpu_ioctl_set_fpu(struct kvm_vcpu *vcpu, struct kvm_fpu *fpu); kvm_vcpu 836 include/linux/kvm_host.h int kvm_arch_vcpu_ioctl_translate(struct kvm_vcpu *vcpu, kvm_vcpu 839 include/linux/kvm_host.h int kvm_arch_vcpu_ioctl_get_regs(struct kvm_vcpu *vcpu, struct kvm_regs *regs); kvm_vcpu 840 include/linux/kvm_host.h int kvm_arch_vcpu_ioctl_set_regs(struct kvm_vcpu *vcpu, struct kvm_regs *regs); kvm_vcpu 841 include/linux/kvm_host.h int kvm_arch_vcpu_ioctl_get_sregs(struct kvm_vcpu *vcpu, kvm_vcpu 843 include/linux/kvm_host.h int kvm_arch_vcpu_ioctl_set_sregs(struct kvm_vcpu *vcpu, kvm_vcpu 845 include/linux/kvm_host.h int kvm_arch_vcpu_ioctl_get_mpstate(struct kvm_vcpu *vcpu, kvm_vcpu 847 include/linux/kvm_host.h int kvm_arch_vcpu_ioctl_set_mpstate(struct kvm_vcpu *vcpu, kvm_vcpu 849 include/linux/kvm_host.h int kvm_arch_vcpu_ioctl_set_guest_debug(struct kvm_vcpu *vcpu, kvm_vcpu 851 include/linux/kvm_host.h int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run); kvm_vcpu 856 include/linux/kvm_host.h int kvm_arch_vcpu_init(struct kvm_vcpu *vcpu); kvm_vcpu 857 include/linux/kvm_host.h void kvm_arch_vcpu_uninit(struct kvm_vcpu *vcpu); kvm_vcpu 859 include/linux/kvm_host.h void kvm_arch_sched_in(struct kvm_vcpu *vcpu, int cpu); kvm_vcpu 861 include/linux/kvm_host.h void kvm_arch_vcpu_free(struct kvm_vcpu *vcpu); kvm_vcpu 862 include/linux/kvm_host.h void kvm_arch_vcpu_load(struct kvm_vcpu *vcpu, int cpu); kvm_vcpu 863 include/linux/kvm_host.h void kvm_arch_vcpu_put(struct kvm_vcpu *vcpu); kvm_vcpu 864 include/linux/kvm_host.h struct kvm_vcpu *kvm_arch_vcpu_create(struct kvm *kvm, unsigned int id); kvm_vcpu 865 include/linux/kvm_host.h int kvm_arch_vcpu_setup(struct kvm_vcpu *vcpu); kvm_vcpu 866 include/linux/kvm_host.h void kvm_arch_vcpu_postcreate(struct kvm_vcpu *vcpu); kvm_vcpu 867 include/linux/kvm_host.h void kvm_arch_vcpu_destroy(struct kvm_vcpu *vcpu); kvm_vcpu 870 include/linux/kvm_host.h void kvm_arch_create_vcpu_debugfs(struct kvm_vcpu *vcpu); kvm_vcpu 878 include/linux/kvm_host.h int kvm_arch_vcpu_runnable(struct kvm_vcpu *vcpu); kvm_vcpu 879 include/linux/kvm_host.h bool kvm_arch_vcpu_in_kernel(struct kvm_vcpu *vcpu); kvm_vcpu 880 include/linux/kvm_host.h int kvm_arch_vcpu_should_kick(struct kvm_vcpu *vcpu); kvm_vcpu 881 include/linux/kvm_host.h bool kvm_arch_dy_runnable(struct kvm_vcpu *vcpu); kvm_vcpu 943 include/linux/kvm_host.h static inline struct swait_queue_head *kvm_arch_vcpu_wq(struct kvm_vcpu *vcpu) kvm_vcpu 970 include/linux/kvm_host.h int kvm_cpu_has_pending_timer(struct kvm_vcpu *vcpu); kvm_vcpu 971 include/linux/kvm_host.h void kvm_vcpu_kick(struct kvm_vcpu *vcpu); kvm_vcpu 1079 include/linux/kvm_host.h static inline struct page *kvm_vcpu_gpa_to_page(struct kvm_vcpu *vcpu, kvm_vcpu 1200 include/linux/kvm_host.h static inline void kvm_make_request(int req, struct kvm_vcpu *vcpu) kvm_vcpu 1210 include/linux/kvm_host.h static inline bool kvm_request_pending(struct kvm_vcpu *vcpu) kvm_vcpu 1215 include/linux/kvm_host.h static inline bool kvm_test_request(int req, struct kvm_vcpu *vcpu) kvm_vcpu 1220 include/linux/kvm_host.h static inline void kvm_clear_request(int req, struct kvm_vcpu *vcpu) kvm_vcpu 1225 include/linux/kvm_host.h static inline bool kvm_check_request(int req, struct kvm_vcpu *vcpu) kvm_vcpu 1311 include/linux/kvm_host.h static inline void kvm_vcpu_set_in_spin_loop(struct kvm_vcpu *vcpu, bool val) kvm_vcpu 1315 include/linux/kvm_host.h static inline void kvm_vcpu_set_dy_eligible(struct kvm_vcpu *vcpu, bool val) kvm_vcpu 1322 include/linux/kvm_host.h static inline void kvm_vcpu_set_in_spin_loop(struct kvm_vcpu *vcpu, bool val) kvm_vcpu 1326 include/linux/kvm_host.h static inline void kvm_vcpu_set_dy_eligible(struct kvm_vcpu *vcpu, bool val) kvm_vcpu 1345 include/linux/kvm_host.h static inline bool vcpu_valid_wakeup(struct kvm_vcpu *vcpu) kvm_vcpu 1351 include/linux/kvm_host.h static inline bool vcpu_valid_wakeup(struct kvm_vcpu *vcpu) kvm_vcpu 1359 include/linux/kvm_host.h bool kvm_arch_no_poll(struct kvm_vcpu *vcpu); kvm_vcpu 1361 include/linux/kvm_host.h static inline bool kvm_arch_no_poll(struct kvm_vcpu *vcpu) kvm_vcpu 1383 include/linux/kvm_host.h int kvm_arch_vcpu_run_pid_change(struct kvm_vcpu *vcpu); kvm_vcpu 1385 include/linux/kvm_host.h static inline int kvm_arch_vcpu_run_pid_change(struct kvm_vcpu *vcpu) kvm_vcpu 15 include/linux/kvm_types.h struct kvm_vcpu; kvm_vcpu 36 virt/kvm/arm/aarch32.c static bool pre_fault_synchronize(struct kvm_vcpu *vcpu) kvm_vcpu 48 virt/kvm/arm/aarch32.c static void post_fault_synchronize(struct kvm_vcpu *vcpu, bool loaded) kvm_vcpu 74 virt/kvm/arm/aarch32.c static unsigned long get_except32_cpsr(struct kvm_vcpu *vcpu, u32 mode) kvm_vcpu 154 virt/kvm/arm/aarch32.c static void prepare_fault32(struct kvm_vcpu *vcpu, u32 mode, u32 vect_offset) kvm_vcpu 176 virt/kvm/arm/aarch32.c void kvm_inject_undef32(struct kvm_vcpu *vcpu) kvm_vcpu 188 virt/kvm/arm/aarch32.c static void inject_abt32(struct kvm_vcpu *vcpu, bool is_pabt, kvm_vcpu 224 virt/kvm/arm/aarch32.c void kvm_inject_dabt32(struct kvm_vcpu *vcpu, unsigned long addr) kvm_vcpu 229 virt/kvm/arm/aarch32.c void kvm_inject_pabt32(struct kvm_vcpu *vcpu, unsigned long addr) kvm_vcpu 43 virt/kvm/arm/arch_timer.c static void kvm_timer_update_irq(struct kvm_vcpu *vcpu, bool new_level, kvm_vcpu 46 virt/kvm/arm/arch_timer.c static void kvm_arm_timer_write(struct kvm_vcpu *vcpu, kvm_vcpu 50 virt/kvm/arm/arch_timer.c static u64 kvm_arm_timer_read(struct kvm_vcpu *vcpu, kvm_vcpu 59 virt/kvm/arm/arch_timer.c static void get_timer_map(struct kvm_vcpu *vcpu, struct timer_map *map) kvm_vcpu 93 virt/kvm/arm/arch_timer.c struct kvm_vcpu *vcpu = *(struct kvm_vcpu **)dev_id; kvm_vcpu 155 virt/kvm/arm/arch_timer.c static u64 kvm_timer_earliest_exp(struct kvm_vcpu *vcpu) kvm_vcpu 178 virt/kvm/arm/arch_timer.c struct kvm_vcpu *vcpu; kvm_vcpu 182 virt/kvm/arm/arch_timer.c vcpu = container_of(timer, struct kvm_vcpu, arch.timer_cpu); kvm_vcpu 202 virt/kvm/arm/arch_timer.c struct kvm_vcpu *vcpu; kvm_vcpu 265 virt/kvm/arm/arch_timer.c bool kvm_timer_is_pending(struct kvm_vcpu *vcpu) kvm_vcpu 279 virt/kvm/arm/arch_timer.c void kvm_timer_update_run(struct kvm_vcpu *vcpu) kvm_vcpu 294 virt/kvm/arm/arch_timer.c static void kvm_timer_update_irq(struct kvm_vcpu *vcpu, bool new_level, kvm_vcpu 386 virt/kvm/arm/arch_timer.c static void kvm_timer_blocking(struct kvm_vcpu *vcpu) kvm_vcpu 409 virt/kvm/arm/arch_timer.c static void kvm_timer_unblocking(struct kvm_vcpu *vcpu) kvm_vcpu 476 virt/kvm/arm/arch_timer.c struct kvm_vcpu *vcpu = ctx->vcpu; kvm_vcpu 495 virt/kvm/arm/arch_timer.c static void kvm_timer_vcpu_load_nogic(struct kvm_vcpu *vcpu) kvm_vcpu 523 virt/kvm/arm/arch_timer.c void kvm_timer_vcpu_load(struct kvm_vcpu *vcpu) kvm_vcpu 553 virt/kvm/arm/arch_timer.c bool kvm_timer_should_notify_user(struct kvm_vcpu *vcpu) kvm_vcpu 570 virt/kvm/arm/arch_timer.c void kvm_timer_vcpu_put(struct kvm_vcpu *vcpu) kvm_vcpu 614 virt/kvm/arm/arch_timer.c static void unmask_vtimer_irq_user(struct kvm_vcpu *vcpu) kvm_vcpu 627 virt/kvm/arm/arch_timer.c void kvm_timer_sync_hwstate(struct kvm_vcpu *vcpu) kvm_vcpu 638 virt/kvm/arm/arch_timer.c int kvm_timer_vcpu_reset(struct kvm_vcpu *vcpu) kvm_vcpu 672 virt/kvm/arm/arch_timer.c static void update_vtimer_cntvoff(struct kvm_vcpu *vcpu, u64 cntvoff) kvm_vcpu 676 virt/kvm/arm/arch_timer.c struct kvm_vcpu *tmp; kvm_vcpu 690 virt/kvm/arm/arch_timer.c void kvm_timer_vcpu_init(struct kvm_vcpu *vcpu) kvm_vcpu 727 virt/kvm/arm/arch_timer.c int kvm_arm_timer_set_reg(struct kvm_vcpu *vcpu, u64 regid, u64 value) kvm_vcpu 774 virt/kvm/arm/arch_timer.c u64 kvm_arm_timer_get_reg(struct kvm_vcpu *vcpu, u64 regid) kvm_vcpu 799 virt/kvm/arm/arch_timer.c static u64 kvm_arm_timer_read(struct kvm_vcpu *vcpu, kvm_vcpu 830 virt/kvm/arm/arch_timer.c u64 kvm_arm_timer_read_sysreg(struct kvm_vcpu *vcpu, kvm_vcpu 847 virt/kvm/arm/arch_timer.c static void kvm_arm_timer_write(struct kvm_vcpu *vcpu, kvm_vcpu 870 virt/kvm/arm/arch_timer.c void kvm_arm_timer_write_sysreg(struct kvm_vcpu *vcpu, kvm_vcpu 993 virt/kvm/arm/arch_timer.c void kvm_timer_vcpu_terminate(struct kvm_vcpu *vcpu) kvm_vcpu 1000 virt/kvm/arm/arch_timer.c static bool timer_irqs_are_valid(struct kvm_vcpu *vcpu) kvm_vcpu 1026 virt/kvm/arm/arch_timer.c struct kvm_vcpu *vcpu = kvm_arm_get_running_vcpu(); kvm_vcpu 1039 virt/kvm/arm/arch_timer.c int kvm_timer_enable(struct kvm_vcpu *vcpu) kvm_vcpu 1108 virt/kvm/arm/arch_timer.c struct kvm_vcpu *vcpu; kvm_vcpu 1117 virt/kvm/arm/arch_timer.c int kvm_arm_timer_set_attr(struct kvm_vcpu *vcpu, struct kvm_device_attr *attr) kvm_vcpu 1150 virt/kvm/arm/arch_timer.c int kvm_arm_timer_get_attr(struct kvm_vcpu *vcpu, struct kvm_device_attr *attr) kvm_vcpu 1171 virt/kvm/arm/arch_timer.c int kvm_arm_timer_has_attr(struct kvm_vcpu *vcpu, struct kvm_device_attr *attr) kvm_vcpu 51 virt/kvm/arm/arm.c static DEFINE_PER_CPU(struct kvm_vcpu *, kvm_arm_running_vcpu); kvm_vcpu 62 virt/kvm/arm/arm.c static void kvm_arm_set_running_vcpu(struct kvm_vcpu *vcpu) kvm_vcpu 73 virt/kvm/arm/arm.c struct kvm_vcpu *kvm_arm_get_running_vcpu(void) kvm_vcpu 81 virt/kvm/arm/arm.c struct kvm_vcpu * __percpu *kvm_get_running_vcpus(void) kvm_vcpu 86 virt/kvm/arm/arm.c int kvm_arch_vcpu_should_kick(struct kvm_vcpu *vcpu) kvm_vcpu 147 virt/kvm/arm/arm.c int kvm_arch_create_vcpu_debugfs(struct kvm_vcpu *vcpu) kvm_vcpu 152 virt/kvm/arm/arm.c vm_fault_t kvm_arch_vcpu_fault(struct kvm_vcpu *vcpu, struct vm_fault *vmf) kvm_vcpu 256 virt/kvm/arm/arm.c struct kvm_vcpu *kvm_arch_vcpu_create(struct kvm *kvm, unsigned int id) kvm_vcpu 259 virt/kvm/arm/arm.c struct kvm_vcpu *vcpu; kvm_vcpu 294 virt/kvm/arm/arm.c void kvm_arch_vcpu_postcreate(struct kvm_vcpu *vcpu) kvm_vcpu 298 virt/kvm/arm/arm.c void kvm_arch_vcpu_free(struct kvm_vcpu *vcpu) kvm_vcpu 310 virt/kvm/arm/arm.c void kvm_arch_vcpu_destroy(struct kvm_vcpu *vcpu) kvm_vcpu 315 virt/kvm/arm/arm.c int kvm_cpu_has_pending_timer(struct kvm_vcpu *vcpu) kvm_vcpu 320 virt/kvm/arm/arm.c void kvm_arch_vcpu_blocking(struct kvm_vcpu *vcpu) kvm_vcpu 336 virt/kvm/arm/arm.c void kvm_arch_vcpu_unblocking(struct kvm_vcpu *vcpu) kvm_vcpu 341 virt/kvm/arm/arm.c int kvm_arch_vcpu_init(struct kvm_vcpu *vcpu) kvm_vcpu 367 virt/kvm/arm/arm.c void kvm_arch_vcpu_load(struct kvm_vcpu *vcpu, int cpu) kvm_vcpu 412 virt/kvm/arm/arm.c void kvm_arch_vcpu_put(struct kvm_vcpu *vcpu) kvm_vcpu 425 virt/kvm/arm/arm.c static void vcpu_power_off(struct kvm_vcpu *vcpu) kvm_vcpu 432 virt/kvm/arm/arm.c int kvm_arch_vcpu_ioctl_get_mpstate(struct kvm_vcpu *vcpu, kvm_vcpu 443 virt/kvm/arm/arm.c int kvm_arch_vcpu_ioctl_set_mpstate(struct kvm_vcpu *vcpu, kvm_vcpu 469 virt/kvm/arm/arm.c int kvm_arch_vcpu_runnable(struct kvm_vcpu *v) kvm_vcpu 476 virt/kvm/arm/arm.c bool kvm_arch_vcpu_in_kernel(struct kvm_vcpu *vcpu) kvm_vcpu 563 virt/kvm/arm/arm.c static int kvm_vcpu_first_run_init(struct kvm_vcpu *vcpu) kvm_vcpu 611 virt/kvm/arm/arm.c struct kvm_vcpu *vcpu; kvm_vcpu 621 virt/kvm/arm/arm.c struct kvm_vcpu *vcpu; kvm_vcpu 629 virt/kvm/arm/arm.c static void vcpu_req_sleep(struct kvm_vcpu *vcpu) kvm_vcpu 649 virt/kvm/arm/arm.c static int kvm_vcpu_initialized(struct kvm_vcpu *vcpu) kvm_vcpu 654 virt/kvm/arm/arm.c static void check_vcpu_requests(struct kvm_vcpu *vcpu) kvm_vcpu 682 virt/kvm/arm/arm.c int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu, struct kvm_run *run) kvm_vcpu 867 virt/kvm/arm/arm.c static int vcpu_interrupt_line(struct kvm_vcpu *vcpu, int number, bool level) kvm_vcpu 907 virt/kvm/arm/arm.c struct kvm_vcpu *vcpu = NULL; kvm_vcpu 961 virt/kvm/arm/arm.c static int kvm_vcpu_set_target(struct kvm_vcpu *vcpu, kvm_vcpu 1008 virt/kvm/arm/arm.c static int kvm_arch_vcpu_ioctl_vcpu_init(struct kvm_vcpu *vcpu, kvm_vcpu 1037 virt/kvm/arm/arm.c static int kvm_arm_vcpu_set_attr(struct kvm_vcpu *vcpu, kvm_vcpu 1051 virt/kvm/arm/arm.c static int kvm_arm_vcpu_get_attr(struct kvm_vcpu *vcpu, kvm_vcpu 1065 virt/kvm/arm/arm.c static int kvm_arm_vcpu_has_attr(struct kvm_vcpu *vcpu, kvm_vcpu 1079 virt/kvm/arm/arm.c static int kvm_arm_vcpu_get_events(struct kvm_vcpu *vcpu, kvm_vcpu 1087 virt/kvm/arm/arm.c static int kvm_arm_vcpu_set_events(struct kvm_vcpu *vcpu, kvm_vcpu 1108 virt/kvm/arm/arm.c struct kvm_vcpu *vcpu = filp->private_data; kvm_vcpu 1629 virt/kvm/arm/arm.c struct kvm_vcpu *kvm_mpidr_to_vcpu(struct kvm *kvm, unsigned long mpidr) kvm_vcpu 1631 virt/kvm/arm/arm.c struct kvm_vcpu *vcpu; kvm_vcpu 1751 virt/kvm/arm/arm.c int rc = kvm_init(NULL, sizeof(struct kvm_vcpu), 0, THIS_MODULE); kvm_vcpu 47 virt/kvm/arm/hyp/aarch32.c bool __hyp_text kvm_condition_valid32(const struct kvm_vcpu *vcpu) kvm_vcpu 96 virt/kvm/arm/hyp/aarch32.c static void __hyp_text kvm_adjust_itstate(struct kvm_vcpu *vcpu) kvm_vcpu 126 virt/kvm/arm/hyp/aarch32.c void __hyp_text kvm_skip_instr32(struct kvm_vcpu *vcpu, bool is_wide_instr) kvm_vcpu 23 virt/kvm/arm/hyp/timer-sr.c void __hyp_text __timer_disable_traps(struct kvm_vcpu *vcpu) kvm_vcpu 37 virt/kvm/arm/hyp/timer-sr.c void __hyp_text __timer_enable_traps(struct kvm_vcpu *vcpu) kvm_vcpu 197 virt/kvm/arm/hyp/vgic-v3-sr.c void __hyp_text __vgic_v3_save_state(struct kvm_vcpu *vcpu) kvm_vcpu 233 virt/kvm/arm/hyp/vgic-v3-sr.c void __hyp_text __vgic_v3_restore_state(struct kvm_vcpu *vcpu) kvm_vcpu 260 virt/kvm/arm/hyp/vgic-v3-sr.c void __hyp_text __vgic_v3_activate_traps(struct kvm_vcpu *vcpu) kvm_vcpu 309 virt/kvm/arm/hyp/vgic-v3-sr.c void __hyp_text __vgic_v3_deactivate_traps(struct kvm_vcpu *vcpu) kvm_vcpu 336 virt/kvm/arm/hyp/vgic-v3-sr.c void __hyp_text __vgic_v3_save_aprs(struct kvm_vcpu *vcpu) kvm_vcpu 373 virt/kvm/arm/hyp/vgic-v3-sr.c void __hyp_text __vgic_v3_restore_aprs(struct kvm_vcpu *vcpu) kvm_vcpu 442 virt/kvm/arm/hyp/vgic-v3-sr.c static int __hyp_text __vgic_v3_get_group(struct kvm_vcpu *vcpu) kvm_vcpu 452 virt/kvm/arm/hyp/vgic-v3-sr.c static int __hyp_text __vgic_v3_highest_priority_lr(struct kvm_vcpu *vcpu, kvm_vcpu 492 virt/kvm/arm/hyp/vgic-v3-sr.c static int __hyp_text __vgic_v3_find_active_lr(struct kvm_vcpu *vcpu, kvm_vcpu 643 virt/kvm/arm/hyp/vgic-v3-sr.c static void __hyp_text __vgic_v3_read_iar(struct kvm_vcpu *vcpu, u32 vmcr, int rt) kvm_vcpu 701 virt/kvm/arm/hyp/vgic-v3-sr.c static void __hyp_text __vgic_v3_write_dir(struct kvm_vcpu *vcpu, kvm_vcpu 725 virt/kvm/arm/hyp/vgic-v3-sr.c static void __hyp_text __vgic_v3_write_eoir(struct kvm_vcpu *vcpu, u32 vmcr, int rt) kvm_vcpu 762 virt/kvm/arm/hyp/vgic-v3-sr.c static void __hyp_text __vgic_v3_read_igrpen0(struct kvm_vcpu *vcpu, u32 vmcr, int rt) kvm_vcpu 767 virt/kvm/arm/hyp/vgic-v3-sr.c static void __hyp_text __vgic_v3_read_igrpen1(struct kvm_vcpu *vcpu, u32 vmcr, int rt) kvm_vcpu 772 virt/kvm/arm/hyp/vgic-v3-sr.c static void __hyp_text __vgic_v3_write_igrpen0(struct kvm_vcpu *vcpu, u32 vmcr, int rt) kvm_vcpu 784 virt/kvm/arm/hyp/vgic-v3-sr.c static void __hyp_text __vgic_v3_write_igrpen1(struct kvm_vcpu *vcpu, u32 vmcr, int rt) kvm_vcpu 796 virt/kvm/arm/hyp/vgic-v3-sr.c static void __hyp_text __vgic_v3_read_bpr0(struct kvm_vcpu *vcpu, u32 vmcr, int rt) kvm_vcpu 801 virt/kvm/arm/hyp/vgic-v3-sr.c static void __hyp_text __vgic_v3_read_bpr1(struct kvm_vcpu *vcpu, u32 vmcr, int rt) kvm_vcpu 806 virt/kvm/arm/hyp/vgic-v3-sr.c static void __hyp_text __vgic_v3_write_bpr0(struct kvm_vcpu *vcpu, u32 vmcr, int rt) kvm_vcpu 823 virt/kvm/arm/hyp/vgic-v3-sr.c static void __hyp_text __vgic_v3_write_bpr1(struct kvm_vcpu *vcpu, u32 vmcr, int rt) kvm_vcpu 843 virt/kvm/arm/hyp/vgic-v3-sr.c static void __hyp_text __vgic_v3_read_apxrn(struct kvm_vcpu *vcpu, int rt, int n) kvm_vcpu 855 virt/kvm/arm/hyp/vgic-v3-sr.c static void __hyp_text __vgic_v3_write_apxrn(struct kvm_vcpu *vcpu, int rt, int n) kvm_vcpu 865 virt/kvm/arm/hyp/vgic-v3-sr.c static void __hyp_text __vgic_v3_read_apxr0(struct kvm_vcpu *vcpu, kvm_vcpu 871 virt/kvm/arm/hyp/vgic-v3-sr.c static void __hyp_text __vgic_v3_read_apxr1(struct kvm_vcpu *vcpu, kvm_vcpu 877 virt/kvm/arm/hyp/vgic-v3-sr.c static void __hyp_text __vgic_v3_read_apxr2(struct kvm_vcpu *vcpu, kvm_vcpu 883 virt/kvm/arm/hyp/vgic-v3-sr.c static void __hyp_text __vgic_v3_read_apxr3(struct kvm_vcpu *vcpu, kvm_vcpu 889 virt/kvm/arm/hyp/vgic-v3-sr.c static void __hyp_text __vgic_v3_write_apxr0(struct kvm_vcpu *vcpu, kvm_vcpu 895 virt/kvm/arm/hyp/vgic-v3-sr.c static void __hyp_text __vgic_v3_write_apxr1(struct kvm_vcpu *vcpu, kvm_vcpu 901 virt/kvm/arm/hyp/vgic-v3-sr.c static void __hyp_text __vgic_v3_write_apxr2(struct kvm_vcpu *vcpu, kvm_vcpu 907 virt/kvm/arm/hyp/vgic-v3-sr.c static void __hyp_text __vgic_v3_write_apxr3(struct kvm_vcpu *vcpu, kvm_vcpu 913 virt/kvm/arm/hyp/vgic-v3-sr.c static void __hyp_text __vgic_v3_read_hppir(struct kvm_vcpu *vcpu, kvm_vcpu 933 virt/kvm/arm/hyp/vgic-v3-sr.c static void __hyp_text __vgic_v3_read_pmr(struct kvm_vcpu *vcpu, kvm_vcpu 941 virt/kvm/arm/hyp/vgic-v3-sr.c static void __hyp_text __vgic_v3_write_pmr(struct kvm_vcpu *vcpu, kvm_vcpu 954 virt/kvm/arm/hyp/vgic-v3-sr.c static void __hyp_text __vgic_v3_read_rpr(struct kvm_vcpu *vcpu, kvm_vcpu 961 virt/kvm/arm/hyp/vgic-v3-sr.c static void __hyp_text __vgic_v3_read_ctlr(struct kvm_vcpu *vcpu, kvm_vcpu 983 virt/kvm/arm/hyp/vgic-v3-sr.c static void __hyp_text __vgic_v3_write_ctlr(struct kvm_vcpu *vcpu, kvm_vcpu 1001 virt/kvm/arm/hyp/vgic-v3-sr.c int __hyp_text __vgic_v3_perform_cpuif_access(struct kvm_vcpu *vcpu) kvm_vcpu 1006 virt/kvm/arm/hyp/vgic-v3-sr.c void (*fn)(struct kvm_vcpu *, u32, int); kvm_vcpu 83 virt/kvm/arm/mmio.c int kvm_handle_mmio_return(struct kvm_vcpu *vcpu, struct kvm_run *run) kvm_vcpu 126 virt/kvm/arm/mmio.c static int decode_hsr(struct kvm_vcpu *vcpu, bool *is_write, int *len) kvm_vcpu 156 virt/kvm/arm/mmio.c int io_mem_abort(struct kvm_vcpu *vcpu, struct kvm_run *run, kvm_vcpu 1669 virt/kvm/arm/mmu.c static int user_mem_abort(struct kvm_vcpu *vcpu, phys_addr_t fault_ipa, kvm_vcpu 1861 virt/kvm/arm/mmu.c static void handle_access_fault(struct kvm_vcpu *vcpu, phys_addr_t fault_ipa) kvm_vcpu 1908 virt/kvm/arm/mmu.c int kvm_handle_guest_abort(struct kvm_vcpu *vcpu, struct kvm_run *run) kvm_vcpu 2154 virt/kvm/arm/mmu.c void kvm_mmu_free_memory_caches(struct kvm_vcpu *vcpu) kvm_vcpu 2429 virt/kvm/arm/mmu.c void kvm_set_way_flush(struct kvm_vcpu *vcpu) kvm_vcpu 2450 virt/kvm/arm/mmu.c void kvm_toggle_cache(struct kvm_vcpu *vcpu, bool was_enabled) kvm_vcpu 21 virt/kvm/arm/perf.c struct kvm_vcpu *vcpu; kvm_vcpu 33 virt/kvm/arm/perf.c struct kvm_vcpu *vcpu; kvm_vcpu 17 virt/kvm/arm/pmu.c static void kvm_pmu_create_perf_event(struct kvm_vcpu *vcpu, u64 select_idx); kvm_vcpu 26 virt/kvm/arm/pmu.c static bool kvm_pmu_idx_is_64bit(struct kvm_vcpu *vcpu, u64 select_idx) kvm_vcpu 32 virt/kvm/arm/pmu.c static struct kvm_vcpu *kvm_pmc_to_vcpu(struct kvm_pmc *pmc) kvm_vcpu 40 virt/kvm/arm/pmu.c return container_of(vcpu_arch, struct kvm_vcpu, arch); kvm_vcpu 49 virt/kvm/arm/pmu.c struct kvm_vcpu *vcpu = kvm_pmc_to_vcpu(pmc); kvm_vcpu 84 virt/kvm/arm/pmu.c static bool kvm_pmu_idx_has_chain_evtype(struct kvm_vcpu *vcpu, u64 select_idx) kvm_vcpu 104 virt/kvm/arm/pmu.c static u64 kvm_pmu_get_pair_counter_value(struct kvm_vcpu *vcpu, kvm_vcpu 139 virt/kvm/arm/pmu.c u64 kvm_pmu_get_counter_value(struct kvm_vcpu *vcpu, u64 select_idx) kvm_vcpu 162 virt/kvm/arm/pmu.c void kvm_pmu_set_counter_value(struct kvm_vcpu *vcpu, u64 select_idx, u64 val) kvm_vcpu 194 virt/kvm/arm/pmu.c static void kvm_pmu_stop_counter(struct kvm_vcpu *vcpu, struct kvm_pmc *pmc) kvm_vcpu 225 virt/kvm/arm/pmu.c void kvm_pmu_vcpu_init(struct kvm_vcpu *vcpu) kvm_vcpu 239 virt/kvm/arm/pmu.c void kvm_pmu_vcpu_reset(struct kvm_vcpu *vcpu) kvm_vcpu 255 virt/kvm/arm/pmu.c void kvm_pmu_vcpu_destroy(struct kvm_vcpu *vcpu) kvm_vcpu 264 virt/kvm/arm/pmu.c u64 kvm_pmu_valid_counter_mask(struct kvm_vcpu *vcpu) kvm_vcpu 282 virt/kvm/arm/pmu.c void kvm_pmu_enable_counter_mask(struct kvm_vcpu *vcpu, u64 val) kvm_vcpu 323 virt/kvm/arm/pmu.c void kvm_pmu_disable_counter_mask(struct kvm_vcpu *vcpu, u64 val) kvm_vcpu 354 virt/kvm/arm/pmu.c static u64 kvm_pmu_overflow_status(struct kvm_vcpu *vcpu) kvm_vcpu 368 virt/kvm/arm/pmu.c static void kvm_pmu_update_state(struct kvm_vcpu *vcpu) kvm_vcpu 389 virt/kvm/arm/pmu.c bool kvm_pmu_should_notify_user(struct kvm_vcpu *vcpu) kvm_vcpu 404 virt/kvm/arm/pmu.c void kvm_pmu_update_run(struct kvm_vcpu *vcpu) kvm_vcpu 421 virt/kvm/arm/pmu.c void kvm_pmu_flush_hwstate(struct kvm_vcpu *vcpu) kvm_vcpu 433 virt/kvm/arm/pmu.c void kvm_pmu_sync_hwstate(struct kvm_vcpu *vcpu) kvm_vcpu 447 virt/kvm/arm/pmu.c struct kvm_vcpu *vcpu = kvm_pmc_to_vcpu(pmc); kvm_vcpu 481 virt/kvm/arm/pmu.c void kvm_pmu_software_increment(struct kvm_vcpu *vcpu, u64 val) kvm_vcpu 531 virt/kvm/arm/pmu.c void kvm_pmu_handle_pmcr(struct kvm_vcpu *vcpu, u64 val) kvm_vcpu 553 virt/kvm/arm/pmu.c static bool kvm_pmu_counter_is_enabled(struct kvm_vcpu *vcpu, u64 select_idx) kvm_vcpu 564 virt/kvm/arm/pmu.c static void kvm_pmu_create_perf_event(struct kvm_vcpu *vcpu, u64 select_idx) kvm_vcpu 646 virt/kvm/arm/pmu.c static void kvm_pmu_update_pmc_chained(struct kvm_vcpu *vcpu, u64 select_idx) kvm_vcpu 675 virt/kvm/arm/pmu.c void kvm_pmu_set_counter_event_type(struct kvm_vcpu *vcpu, u64 data, kvm_vcpu 699 virt/kvm/arm/pmu.c int kvm_arm_pmu_v3_enable(struct kvm_vcpu *vcpu) kvm_vcpu 732 virt/kvm/arm/pmu.c static int kvm_arm_pmu_v3_init(struct kvm_vcpu *vcpu) kvm_vcpu 775 virt/kvm/arm/pmu.c struct kvm_vcpu *vcpu; kvm_vcpu 793 virt/kvm/arm/pmu.c int kvm_arm_pmu_v3_set_attr(struct kvm_vcpu *vcpu, struct kvm_device_attr *attr) kvm_vcpu 830 virt/kvm/arm/pmu.c int kvm_arm_pmu_v3_get_attr(struct kvm_vcpu *vcpu, struct kvm_device_attr *attr) kvm_vcpu 854 virt/kvm/arm/pmu.c int kvm_arm_pmu_v3_has_attr(struct kvm_vcpu *vcpu, struct kvm_device_attr *attr) kvm_vcpu 26 virt/kvm/arm/psci.c static u32 smccc_get_function(struct kvm_vcpu *vcpu) kvm_vcpu 31 virt/kvm/arm/psci.c static unsigned long smccc_get_arg1(struct kvm_vcpu *vcpu) kvm_vcpu 36 virt/kvm/arm/psci.c static unsigned long smccc_get_arg2(struct kvm_vcpu *vcpu) kvm_vcpu 41 virt/kvm/arm/psci.c static unsigned long smccc_get_arg3(struct kvm_vcpu *vcpu) kvm_vcpu 46 virt/kvm/arm/psci.c static void smccc_set_retval(struct kvm_vcpu *vcpu, kvm_vcpu 66 virt/kvm/arm/psci.c static unsigned long kvm_psci_vcpu_suspend(struct kvm_vcpu *vcpu) kvm_vcpu 87 virt/kvm/arm/psci.c static void kvm_psci_vcpu_off(struct kvm_vcpu *vcpu) kvm_vcpu 94 virt/kvm/arm/psci.c static unsigned long kvm_psci_vcpu_on(struct kvm_vcpu *source_vcpu) kvm_vcpu 98 virt/kvm/arm/psci.c struct kvm_vcpu *vcpu = NULL; kvm_vcpu 148 virt/kvm/arm/psci.c static unsigned long kvm_psci_vcpu_affinity_info(struct kvm_vcpu *vcpu) kvm_vcpu 156 virt/kvm/arm/psci.c struct kvm_vcpu *tmp; kvm_vcpu 188 virt/kvm/arm/psci.c static void kvm_prepare_system_event(struct kvm_vcpu *vcpu, u32 type) kvm_vcpu 191 virt/kvm/arm/psci.c struct kvm_vcpu *tmp; kvm_vcpu 211 virt/kvm/arm/psci.c static void kvm_psci_system_off(struct kvm_vcpu *vcpu) kvm_vcpu 216 virt/kvm/arm/psci.c static void kvm_psci_system_reset(struct kvm_vcpu *vcpu) kvm_vcpu 221 virt/kvm/arm/psci.c static int kvm_psci_0_2_call(struct kvm_vcpu *vcpu) kvm_vcpu 295 virt/kvm/arm/psci.c static int kvm_psci_1_0_call(struct kvm_vcpu *vcpu) kvm_vcpu 337 virt/kvm/arm/psci.c static int kvm_psci_0_1_call(struct kvm_vcpu *vcpu) kvm_vcpu 376 virt/kvm/arm/psci.c static int kvm_psci_call(struct kvm_vcpu *vcpu) kvm_vcpu 390 virt/kvm/arm/psci.c int kvm_hvc_call_handler(struct kvm_vcpu *vcpu) kvm_vcpu 439 virt/kvm/arm/psci.c int kvm_arm_get_fw_num_regs(struct kvm_vcpu *vcpu) kvm_vcpu 444 virt/kvm/arm/psci.c int kvm_arm_copy_fw_reg_indices(struct kvm_vcpu *vcpu, u64 __user *uindices) kvm_vcpu 496 virt/kvm/arm/psci.c int kvm_arm_get_fw_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg) kvm_vcpu 525 virt/kvm/arm/psci.c int kvm_arm_set_fw_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg) kvm_vcpu 162 virt/kvm/arm/vgic/vgic-debug.c struct kvm_vcpu *vcpu) kvm_vcpu 178 virt/kvm/arm/vgic/vgic-debug.c struct kvm_vcpu *vcpu) kvm_vcpu 223 virt/kvm/arm/vgic/vgic-debug.c struct kvm_vcpu *vcpu = NULL; kvm_vcpu 74 virt/kvm/arm/vgic/vgic-init.c struct kvm_vcpu *vcpu; kvm_vcpu 145 virt/kvm/arm/vgic/vgic-init.c struct kvm_vcpu *vcpu0 = kvm_get_vcpu(kvm, 0); kvm_vcpu 195 virt/kvm/arm/vgic/vgic-init.c int kvm_vgic_vcpu_init(struct kvm_vcpu *vcpu) kvm_vcpu 245 virt/kvm/arm/vgic/vgic-init.c static void kvm_vgic_vcpu_enable(struct kvm_vcpu *vcpu) kvm_vcpu 266 virt/kvm/arm/vgic/vgic-init.c struct kvm_vcpu *vcpu; kvm_vcpu 356 virt/kvm/arm/vgic/vgic-init.c void kvm_vgic_vcpu_destroy(struct kvm_vcpu *vcpu) kvm_vcpu 366 virt/kvm/arm/vgic/vgic-init.c struct kvm_vcpu *vcpu; kvm_vcpu 30 virt/kvm/arm/vgic/vgic-its.c struct kvm_vcpu *filter_vcpu, bool needs_inv); kvm_vcpu 40 virt/kvm/arm/vgic/vgic-its.c struct kvm_vcpu *vcpu) kvm_vcpu 274 virt/kvm/arm/vgic/vgic-its.c struct kvm_vcpu *filter_vcpu, bool needs_inv) kvm_vcpu 312 virt/kvm/arm/vgic/vgic-its.c int vgic_copy_lpi_list(struct kvm *kvm, struct kvm_vcpu *vcpu, u32 **intid_ptr) kvm_vcpu 347 virt/kvm/arm/vgic/vgic-its.c static int update_affinity(struct vgic_irq *irq, struct kvm_vcpu *vcpu) kvm_vcpu 379 virt/kvm/arm/vgic/vgic-its.c struct kvm_vcpu *vcpu; kvm_vcpu 418 virt/kvm/arm/vgic/vgic-its.c static int its_sync_lpi_pending_table(struct kvm_vcpu *vcpu) kvm_vcpu 662 virt/kvm/arm/vgic/vgic-its.c struct kvm_vcpu *vcpu; kvm_vcpu 866 virt/kvm/arm/vgic/vgic-its.c struct kvm_vcpu *vcpu; kvm_vcpu 1046 virt/kvm/arm/vgic/vgic-its.c struct kvm_vcpu *vcpu = NULL; kvm_vcpu 1299 virt/kvm/arm/vgic/vgic-its.c struct kvm_vcpu *vcpu; kvm_vcpu 1343 virt/kvm/arm/vgic/vgic-its.c struct kvm_vcpu *vcpu1, *vcpu2; kvm_vcpu 1790 virt/kvm/arm/vgic/vgic-its.c void vgic_enable_lpis(struct kvm_vcpu *vcpu) kvm_vcpu 2141 virt/kvm/arm/vgic/vgic-its.c struct kvm_vcpu *vcpu = NULL; kvm_vcpu 301 virt/kvm/arm/vgic/vgic-kvm-device.c struct kvm_vcpu *tmp_vcpu; kvm_vcpu 317 virt/kvm/arm/vgic/vgic-kvm-device.c struct kvm_vcpu *tmp_vcpu; kvm_vcpu 350 virt/kvm/arm/vgic/vgic-kvm-device.c struct kvm_vcpu *vcpu; kvm_vcpu 514 virt/kvm/arm/vgic/vgic-kvm-device.c struct kvm_vcpu *vcpu; kvm_vcpu 25 virt/kvm/arm/vgic/vgic-mmio-v2.c static unsigned long vgic_mmio_read_v2_misc(struct kvm_vcpu *vcpu, kvm_vcpu 52 virt/kvm/arm/vgic/vgic-mmio-v2.c static void vgic_mmio_write_v2_misc(struct kvm_vcpu *vcpu, kvm_vcpu 72 virt/kvm/arm/vgic/vgic-mmio-v2.c static int vgic_mmio_uaccess_write_v2_misc(struct kvm_vcpu *vcpu, kvm_vcpu 98 virt/kvm/arm/vgic/vgic-mmio-v2.c static int vgic_mmio_uaccess_write_v2_group(struct kvm_vcpu *vcpu, kvm_vcpu 108 virt/kvm/arm/vgic/vgic-mmio-v2.c static void vgic_mmio_write_sgir(struct kvm_vcpu *source_vcpu, kvm_vcpu 117 virt/kvm/arm/vgic/vgic-mmio-v2.c struct kvm_vcpu *vcpu; kvm_vcpu 151 virt/kvm/arm/vgic/vgic-mmio-v2.c static unsigned long vgic_mmio_read_target(struct kvm_vcpu *vcpu, kvm_vcpu 169 virt/kvm/arm/vgic/vgic-mmio-v2.c static void vgic_mmio_write_target(struct kvm_vcpu *vcpu, kvm_vcpu 197 virt/kvm/arm/vgic/vgic-mmio-v2.c static unsigned long vgic_mmio_read_sgipend(struct kvm_vcpu *vcpu, kvm_vcpu 214 virt/kvm/arm/vgic/vgic-mmio-v2.c static void vgic_mmio_write_sgipendc(struct kvm_vcpu *vcpu, kvm_vcpu 236 virt/kvm/arm/vgic/vgic-mmio-v2.c static void vgic_mmio_write_sgipends(struct kvm_vcpu *vcpu, kvm_vcpu 264 virt/kvm/arm/vgic/vgic-mmio-v2.c static unsigned long vgic_mmio_read_vcpuif(struct kvm_vcpu *vcpu, kvm_vcpu 311 virt/kvm/arm/vgic/vgic-mmio-v2.c static void vgic_mmio_write_vcpuif(struct kvm_vcpu *vcpu, kvm_vcpu 351 virt/kvm/arm/vgic/vgic-mmio-v2.c static unsigned long vgic_mmio_read_apr(struct kvm_vcpu *vcpu, kvm_vcpu 376 virt/kvm/arm/vgic/vgic-mmio-v2.c static void vgic_mmio_write_apr(struct kvm_vcpu *vcpu, kvm_vcpu 487 virt/kvm/arm/vgic/vgic-mmio-v2.c struct kvm_vcpu *vcpu; kvm_vcpu 524 virt/kvm/arm/vgic/vgic-mmio-v2.c int vgic_v2_cpuif_uaccess(struct kvm_vcpu *vcpu, bool is_write, kvm_vcpu 536 virt/kvm/arm/vgic/vgic-mmio-v2.c int vgic_v2_dist_uaccess(struct kvm_vcpu *vcpu, bool is_write, kvm_vcpu 61 virt/kvm/arm/vgic/vgic-mmio-v3.c static unsigned long vgic_mmio_read_v3_misc(struct kvm_vcpu *vcpu, kvm_vcpu 95 virt/kvm/arm/vgic/vgic-mmio-v3.c static void vgic_mmio_write_v3_misc(struct kvm_vcpu *vcpu, kvm_vcpu 115 virt/kvm/arm/vgic/vgic-mmio-v3.c static int vgic_mmio_uaccess_write_v3_misc(struct kvm_vcpu *vcpu, kvm_vcpu 129 virt/kvm/arm/vgic/vgic-mmio-v3.c static unsigned long vgic_mmio_read_irouter(struct kvm_vcpu *vcpu, kvm_vcpu 147 virt/kvm/arm/vgic/vgic-mmio-v3.c static void vgic_mmio_write_irouter(struct kvm_vcpu *vcpu, kvm_vcpu 174 virt/kvm/arm/vgic/vgic-mmio-v3.c static unsigned long vgic_mmio_read_v3r_ctlr(struct kvm_vcpu *vcpu, kvm_vcpu 183 virt/kvm/arm/vgic/vgic-mmio-v3.c static void vgic_mmio_write_v3r_ctlr(struct kvm_vcpu *vcpu, kvm_vcpu 204 virt/kvm/arm/vgic/vgic-mmio-v3.c static unsigned long vgic_mmio_read_v3r_typer(struct kvm_vcpu *vcpu, kvm_vcpu 226 virt/kvm/arm/vgic/vgic-mmio-v3.c static unsigned long vgic_mmio_read_v3r_iidr(struct kvm_vcpu *vcpu, kvm_vcpu 232 virt/kvm/arm/vgic/vgic-mmio-v3.c static unsigned long vgic_mmio_read_v3_idregs(struct kvm_vcpu *vcpu, kvm_vcpu 244 virt/kvm/arm/vgic/vgic-mmio-v3.c static unsigned long vgic_v3_uaccess_read_pending(struct kvm_vcpu *vcpu, kvm_vcpu 270 virt/kvm/arm/vgic/vgic-mmio-v3.c static int vgic_v3_uaccess_write_pending(struct kvm_vcpu *vcpu, kvm_vcpu 384 virt/kvm/arm/vgic/vgic-mmio-v3.c static unsigned long vgic_mmio_read_propbase(struct kvm_vcpu *vcpu, kvm_vcpu 392 virt/kvm/arm/vgic/vgic-mmio-v3.c static void vgic_mmio_write_propbase(struct kvm_vcpu *vcpu, kvm_vcpu 413 virt/kvm/arm/vgic/vgic-mmio-v3.c static unsigned long vgic_mmio_read_pendbase(struct kvm_vcpu *vcpu, kvm_vcpu 421 virt/kvm/arm/vgic/vgic-mmio-v3.c static void vgic_mmio_write_pendbase(struct kvm_vcpu *vcpu, kvm_vcpu 605 virt/kvm/arm/vgic/vgic-mmio-v3.c int vgic_register_redist_iodev(struct kvm_vcpu *vcpu) kvm_vcpu 654 virt/kvm/arm/vgic/vgic-mmio-v3.c static void vgic_unregister_redist_iodev(struct kvm_vcpu *vcpu) kvm_vcpu 663 virt/kvm/arm/vgic/vgic-mmio-v3.c struct kvm_vcpu *vcpu; kvm_vcpu 790 virt/kvm/arm/vgic/vgic-mmio-v3.c struct kvm_vcpu *vcpu; kvm_vcpu 839 virt/kvm/arm/vgic/vgic-mmio-v3.c static int match_mpidr(u64 sgi_aff, u16 sgi_cpu_mask, struct kvm_vcpu *vcpu) kvm_vcpu 886 virt/kvm/arm/vgic/vgic-mmio-v3.c void vgic_v3_dispatch_sgi(struct kvm_vcpu *vcpu, u64 reg, bool allow_group1) kvm_vcpu 889 virt/kvm/arm/vgic/vgic-mmio-v3.c struct kvm_vcpu *c_vcpu; kvm_vcpu 952 virt/kvm/arm/vgic/vgic-mmio-v3.c int vgic_v3_dist_uaccess(struct kvm_vcpu *vcpu, bool is_write, kvm_vcpu 963 virt/kvm/arm/vgic/vgic-mmio-v3.c int vgic_v3_redist_uaccess(struct kvm_vcpu *vcpu, bool is_write, kvm_vcpu 974 virt/kvm/arm/vgic/vgic-mmio-v3.c int vgic_v3_line_level_info_uaccess(struct kvm_vcpu *vcpu, bool is_write, kvm_vcpu 17 virt/kvm/arm/vgic/vgic-mmio.c unsigned long vgic_mmio_read_raz(struct kvm_vcpu *vcpu, kvm_vcpu 23 virt/kvm/arm/vgic/vgic-mmio.c unsigned long vgic_mmio_read_rao(struct kvm_vcpu *vcpu, kvm_vcpu 29 virt/kvm/arm/vgic/vgic-mmio.c void vgic_mmio_write_wi(struct kvm_vcpu *vcpu, gpa_t addr, kvm_vcpu 35 virt/kvm/arm/vgic/vgic-mmio.c int vgic_mmio_uaccess_write_wi(struct kvm_vcpu *vcpu, gpa_t addr, kvm_vcpu 42 virt/kvm/arm/vgic/vgic-mmio.c unsigned long vgic_mmio_read_group(struct kvm_vcpu *vcpu, kvm_vcpu 62 virt/kvm/arm/vgic/vgic-mmio.c void vgic_mmio_write_group(struct kvm_vcpu *vcpu, gpa_t addr, kvm_vcpu 84 virt/kvm/arm/vgic/vgic-mmio.c unsigned long vgic_mmio_read_enable(struct kvm_vcpu *vcpu, kvm_vcpu 104 virt/kvm/arm/vgic/vgic-mmio.c void vgic_mmio_write_senable(struct kvm_vcpu *vcpu, kvm_vcpu 139 virt/kvm/arm/vgic/vgic-mmio.c void vgic_mmio_write_cenable(struct kvm_vcpu *vcpu, kvm_vcpu 159 virt/kvm/arm/vgic/vgic-mmio.c unsigned long vgic_mmio_read_pending(struct kvm_vcpu *vcpu, kvm_vcpu 193 virt/kvm/arm/vgic/vgic-mmio.c static struct kvm_vcpu *vgic_get_mmio_requester_vcpu(void) kvm_vcpu 195 virt/kvm/arm/vgic/vgic-mmio.c struct kvm_vcpu *vcpu; kvm_vcpu 204 virt/kvm/arm/vgic/vgic-mmio.c static void vgic_hw_irq_spending(struct kvm_vcpu *vcpu, struct vgic_irq *irq, kvm_vcpu 214 virt/kvm/arm/vgic/vgic-mmio.c static bool is_vgic_v2_sgi(struct kvm_vcpu *vcpu, struct vgic_irq *irq) kvm_vcpu 220 virt/kvm/arm/vgic/vgic-mmio.c void vgic_mmio_write_spending(struct kvm_vcpu *vcpu, kvm_vcpu 249 virt/kvm/arm/vgic/vgic-mmio.c static void vgic_hw_irq_cpending(struct kvm_vcpu *vcpu, struct vgic_irq *irq, kvm_vcpu 273 virt/kvm/arm/vgic/vgic-mmio.c void vgic_mmio_write_cpending(struct kvm_vcpu *vcpu, kvm_vcpu 319 virt/kvm/arm/vgic/vgic-mmio.c static void vgic_access_active_prepare(struct kvm_vcpu *vcpu, u32 intid) kvm_vcpu 327 virt/kvm/arm/vgic/vgic-mmio.c static void vgic_access_active_finish(struct kvm_vcpu *vcpu, u32 intid) kvm_vcpu 334 virt/kvm/arm/vgic/vgic-mmio.c static unsigned long __vgic_mmio_read_active(struct kvm_vcpu *vcpu, kvm_vcpu 358 virt/kvm/arm/vgic/vgic-mmio.c unsigned long vgic_mmio_read_active(struct kvm_vcpu *vcpu, kvm_vcpu 375 virt/kvm/arm/vgic/vgic-mmio.c unsigned long vgic_uaccess_read_active(struct kvm_vcpu *vcpu, kvm_vcpu 382 virt/kvm/arm/vgic/vgic-mmio.c static void vgic_hw_irq_change_active(struct kvm_vcpu *vcpu, struct vgic_irq *irq, kvm_vcpu 392 virt/kvm/arm/vgic/vgic-mmio.c static void vgic_mmio_change_active(struct kvm_vcpu *vcpu, struct vgic_irq *irq, kvm_vcpu 396 virt/kvm/arm/vgic/vgic-mmio.c struct kvm_vcpu *requester_vcpu = vgic_get_mmio_requester_vcpu(); kvm_vcpu 432 virt/kvm/arm/vgic/vgic-mmio.c static void __vgic_mmio_write_cactive(struct kvm_vcpu *vcpu, kvm_vcpu 446 virt/kvm/arm/vgic/vgic-mmio.c void vgic_mmio_write_cactive(struct kvm_vcpu *vcpu, kvm_vcpu 461 virt/kvm/arm/vgic/vgic-mmio.c int vgic_mmio_uaccess_write_cactive(struct kvm_vcpu *vcpu, kvm_vcpu 469 virt/kvm/arm/vgic/vgic-mmio.c static void __vgic_mmio_write_sactive(struct kvm_vcpu *vcpu, kvm_vcpu 483 virt/kvm/arm/vgic/vgic-mmio.c void vgic_mmio_write_sactive(struct kvm_vcpu *vcpu, kvm_vcpu 498 virt/kvm/arm/vgic/vgic-mmio.c int vgic_mmio_uaccess_write_sactive(struct kvm_vcpu *vcpu, kvm_vcpu 506 virt/kvm/arm/vgic/vgic-mmio.c unsigned long vgic_mmio_read_priority(struct kvm_vcpu *vcpu, kvm_vcpu 531 virt/kvm/arm/vgic/vgic-mmio.c void vgic_mmio_write_priority(struct kvm_vcpu *vcpu, kvm_vcpu 551 virt/kvm/arm/vgic/vgic-mmio.c unsigned long vgic_mmio_read_config(struct kvm_vcpu *vcpu, kvm_vcpu 570 virt/kvm/arm/vgic/vgic-mmio.c void vgic_mmio_write_config(struct kvm_vcpu *vcpu, kvm_vcpu 603 virt/kvm/arm/vgic/vgic-mmio.c u64 vgic_read_irq_line_level_info(struct kvm_vcpu *vcpu, u32 intid) kvm_vcpu 625 virt/kvm/arm/vgic/vgic-mmio.c void vgic_write_irq_line_level_info(struct kvm_vcpu *vcpu, u32 intid, kvm_vcpu 680 virt/kvm/arm/vgic/vgic-mmio.c void vgic_set_vmcr(struct kvm_vcpu *vcpu, struct vgic_vmcr *vmcr) kvm_vcpu 688 virt/kvm/arm/vgic/vgic-mmio.c void vgic_get_vmcr(struct kvm_vcpu *vcpu, struct vgic_vmcr *vmcr) kvm_vcpu 787 virt/kvm/arm/vgic/vgic-mmio.c vgic_get_mmio_region(struct kvm_vcpu *vcpu, struct vgic_io_device *iodev, kvm_vcpu 800 virt/kvm/arm/vgic/vgic-mmio.c static int vgic_uaccess_read(struct kvm_vcpu *vcpu, struct kvm_io_device *dev, kvm_vcpu 805 virt/kvm/arm/vgic/vgic-mmio.c struct kvm_vcpu *r_vcpu; kvm_vcpu 822 virt/kvm/arm/vgic/vgic-mmio.c static int vgic_uaccess_write(struct kvm_vcpu *vcpu, struct kvm_io_device *dev, kvm_vcpu 827 virt/kvm/arm/vgic/vgic-mmio.c struct kvm_vcpu *r_vcpu; kvm_vcpu 844 virt/kvm/arm/vgic/vgic-mmio.c int vgic_uaccess(struct kvm_vcpu *vcpu, struct vgic_io_device *dev, kvm_vcpu 853 virt/kvm/arm/vgic/vgic-mmio.c static int dispatch_mmio_read(struct kvm_vcpu *vcpu, struct kvm_io_device *dev, kvm_vcpu 885 virt/kvm/arm/vgic/vgic-mmio.c static int dispatch_mmio_write(struct kvm_vcpu *vcpu, struct kvm_io_device *dev, kvm_vcpu 14 virt/kvm/arm/vgic/vgic-mmio.h unsigned long (*read)(struct kvm_vcpu *vcpu, gpa_t addr, kvm_vcpu 20 virt/kvm/arm/vgic/vgic-mmio.h void (*write)(struct kvm_vcpu *vcpu, gpa_t addr, kvm_vcpu 26 virt/kvm/arm/vgic/vgic-mmio.h unsigned long (*uaccess_read)(struct kvm_vcpu *vcpu, gpa_t addr, kvm_vcpu 29 virt/kvm/arm/vgic/vgic-mmio.h int (*uaccess_write)(struct kvm_vcpu *vcpu, gpa_t addr, kvm_vcpu 101 virt/kvm/arm/vgic/vgic-mmio.h int kvm_vgic_register_mmio_region(struct kvm *kvm, struct kvm_vcpu *vcpu, kvm_vcpu 117 virt/kvm/arm/vgic/vgic-mmio.h unsigned long vgic_mmio_read_raz(struct kvm_vcpu *vcpu, kvm_vcpu 120 virt/kvm/arm/vgic/vgic-mmio.h unsigned long vgic_mmio_read_rao(struct kvm_vcpu *vcpu, kvm_vcpu 123 virt/kvm/arm/vgic/vgic-mmio.h void vgic_mmio_write_wi(struct kvm_vcpu *vcpu, gpa_t addr, kvm_vcpu 126 virt/kvm/arm/vgic/vgic-mmio.h int vgic_mmio_uaccess_write_wi(struct kvm_vcpu *vcpu, gpa_t addr, kvm_vcpu 129 virt/kvm/arm/vgic/vgic-mmio.h unsigned long vgic_mmio_read_group(struct kvm_vcpu *vcpu, gpa_t addr, kvm_vcpu 132 virt/kvm/arm/vgic/vgic-mmio.h void vgic_mmio_write_group(struct kvm_vcpu *vcpu, gpa_t addr, kvm_vcpu 135 virt/kvm/arm/vgic/vgic-mmio.h unsigned long vgic_mmio_read_enable(struct kvm_vcpu *vcpu, kvm_vcpu 138 virt/kvm/arm/vgic/vgic-mmio.h void vgic_mmio_write_senable(struct kvm_vcpu *vcpu, kvm_vcpu 142 virt/kvm/arm/vgic/vgic-mmio.h void vgic_mmio_write_cenable(struct kvm_vcpu *vcpu, kvm_vcpu 146 virt/kvm/arm/vgic/vgic-mmio.h unsigned long vgic_mmio_read_pending(struct kvm_vcpu *vcpu, kvm_vcpu 149 virt/kvm/arm/vgic/vgic-mmio.h void vgic_mmio_write_spending(struct kvm_vcpu *vcpu, kvm_vcpu 153 virt/kvm/arm/vgic/vgic-mmio.h void vgic_mmio_write_cpending(struct kvm_vcpu *vcpu, kvm_vcpu 157 virt/kvm/arm/vgic/vgic-mmio.h unsigned long vgic_mmio_read_active(struct kvm_vcpu *vcpu, kvm_vcpu 160 virt/kvm/arm/vgic/vgic-mmio.h unsigned long vgic_uaccess_read_active(struct kvm_vcpu *vcpu, kvm_vcpu 163 virt/kvm/arm/vgic/vgic-mmio.h void vgic_mmio_write_cactive(struct kvm_vcpu *vcpu, kvm_vcpu 167 virt/kvm/arm/vgic/vgic-mmio.h void vgic_mmio_write_sactive(struct kvm_vcpu *vcpu, kvm_vcpu 171 virt/kvm/arm/vgic/vgic-mmio.h int vgic_mmio_uaccess_write_cactive(struct kvm_vcpu *vcpu, kvm_vcpu 175 virt/kvm/arm/vgic/vgic-mmio.h int vgic_mmio_uaccess_write_sactive(struct kvm_vcpu *vcpu, kvm_vcpu 179 virt/kvm/arm/vgic/vgic-mmio.h unsigned long vgic_mmio_read_priority(struct kvm_vcpu *vcpu, kvm_vcpu 182 virt/kvm/arm/vgic/vgic-mmio.h void vgic_mmio_write_priority(struct kvm_vcpu *vcpu, kvm_vcpu 186 virt/kvm/arm/vgic/vgic-mmio.h unsigned long vgic_mmio_read_config(struct kvm_vcpu *vcpu, kvm_vcpu 189 virt/kvm/arm/vgic/vgic-mmio.h void vgic_mmio_write_config(struct kvm_vcpu *vcpu, kvm_vcpu 193 virt/kvm/arm/vgic/vgic-mmio.h int vgic_uaccess(struct kvm_vcpu *vcpu, struct vgic_io_device *dev, kvm_vcpu 196 virt/kvm/arm/vgic/vgic-mmio.h u64 vgic_read_irq_line_level_info(struct kvm_vcpu *vcpu, u32 intid); kvm_vcpu 198 virt/kvm/arm/vgic/vgic-mmio.h void vgic_write_irq_line_level_info(struct kvm_vcpu *vcpu, u32 intid, kvm_vcpu 29 virt/kvm/arm/vgic/vgic-v2.c void vgic_v2_set_underflow(struct kvm_vcpu *vcpu) kvm_vcpu 49 virt/kvm/arm/vgic/vgic-v2.c void vgic_v2_fold_lr_state(struct kvm_vcpu *vcpu) kvm_vcpu 137 virt/kvm/arm/vgic/vgic-v2.c void vgic_v2_populate_lr(struct kvm_vcpu *vcpu, struct vgic_irq *irq, int lr) kvm_vcpu 215 virt/kvm/arm/vgic/vgic-v2.c void vgic_v2_clear_lr(struct kvm_vcpu *vcpu, int lr) kvm_vcpu 220 virt/kvm/arm/vgic/vgic-v2.c void vgic_v2_set_vmcr(struct kvm_vcpu *vcpu, struct vgic_vmcr *vmcrp) kvm_vcpu 247 virt/kvm/arm/vgic/vgic-v2.c void vgic_v2_get_vmcr(struct kvm_vcpu *vcpu, struct vgic_vmcr *vmcrp) kvm_vcpu 275 virt/kvm/arm/vgic/vgic-v2.c void vgic_v2_enable(struct kvm_vcpu *vcpu) kvm_vcpu 427 virt/kvm/arm/vgic/vgic-v2.c static void save_lrs(struct kvm_vcpu *vcpu, void __iomem *base) kvm_vcpu 448 virt/kvm/arm/vgic/vgic-v2.c void vgic_v2_save_state(struct kvm_vcpu *vcpu) kvm_vcpu 462 virt/kvm/arm/vgic/vgic-v2.c void vgic_v2_restore_state(struct kvm_vcpu *vcpu) kvm_vcpu 481 virt/kvm/arm/vgic/vgic-v2.c void vgic_v2_load(struct kvm_vcpu *vcpu) kvm_vcpu 491 virt/kvm/arm/vgic/vgic-v2.c void vgic_v2_vmcr_sync(struct kvm_vcpu *vcpu) kvm_vcpu 498 virt/kvm/arm/vgic/vgic-v2.c void vgic_v2_put(struct kvm_vcpu *vcpu) kvm_vcpu 18 virt/kvm/arm/vgic/vgic-v3.c void vgic_v3_set_underflow(struct kvm_vcpu *vcpu) kvm_vcpu 31 virt/kvm/arm/vgic/vgic-v3.c void vgic_v3_fold_lr_state(struct kvm_vcpu *vcpu) kvm_vcpu 118 virt/kvm/arm/vgic/vgic-v3.c void vgic_v3_populate_lr(struct kvm_vcpu *vcpu, struct vgic_irq *irq, int lr) kvm_vcpu 200 virt/kvm/arm/vgic/vgic-v3.c void vgic_v3_clear_lr(struct kvm_vcpu *vcpu, int lr) kvm_vcpu 205 virt/kvm/arm/vgic/vgic-v3.c void vgic_v3_set_vmcr(struct kvm_vcpu *vcpu, struct vgic_vmcr *vmcrp) kvm_vcpu 235 virt/kvm/arm/vgic/vgic-v3.c void vgic_v3_get_vmcr(struct kvm_vcpu *vcpu, struct vgic_vmcr *vmcrp) kvm_vcpu 271 virt/kvm/arm/vgic/vgic-v3.c void vgic_v3_enable(struct kvm_vcpu *vcpu) kvm_vcpu 316 virt/kvm/arm/vgic/vgic-v3.c struct kvm_vcpu *vcpu; kvm_vcpu 373 virt/kvm/arm/vgic/vgic-v3.c struct kvm_vcpu *vcpu; kvm_vcpu 499 virt/kvm/arm/vgic/vgic-v3.c struct kvm_vcpu *vcpu; kvm_vcpu 651 virt/kvm/arm/vgic/vgic-v3.c void vgic_v3_load(struct kvm_vcpu *vcpu) kvm_vcpu 669 virt/kvm/arm/vgic/vgic-v3.c void vgic_v3_vmcr_sync(struct kvm_vcpu *vcpu) kvm_vcpu 677 virt/kvm/arm/vgic/vgic-v3.c void vgic_v3_put(struct kvm_vcpu *vcpu) kvm_vcpu 86 virt/kvm/arm/vgic/vgic-v4.c struct kvm_vcpu *vcpu = info; kvm_vcpu 107 virt/kvm/arm/vgic/vgic-v4.c struct kvm_vcpu *vcpu; kvm_vcpu 182 virt/kvm/arm/vgic/vgic-v4.c struct kvm_vcpu *vcpu = kvm_get_vcpu(kvm, i); kvm_vcpu 195 virt/kvm/arm/vgic/vgic-v4.c int vgic_v4_sync_hwstate(struct kvm_vcpu *vcpu) kvm_vcpu 203 virt/kvm/arm/vgic/vgic-v4.c int vgic_v4_flush_hwstate(struct kvm_vcpu *vcpu) kvm_vcpu 339 virt/kvm/arm/vgic/vgic-v4.c void kvm_vgic_v4_enable_doorbell(struct kvm_vcpu *vcpu) kvm_vcpu 348 virt/kvm/arm/vgic/vgic-v4.c void kvm_vgic_v4_disable_doorbell(struct kvm_vcpu *vcpu) kvm_vcpu 90 virt/kvm/arm/vgic/vgic.c struct vgic_irq *vgic_get_irq(struct kvm *kvm, struct kvm_vcpu *vcpu, kvm_vcpu 151 virt/kvm/arm/vgic/vgic.c void vgic_flush_pending_lpis(struct kvm_vcpu *vcpu) kvm_vcpu 215 virt/kvm/arm/vgic/vgic.c static struct kvm_vcpu *vgic_target_oracle(struct vgic_irq *irq) kvm_vcpu 297 virt/kvm/arm/vgic/vgic.c static void vgic_sort_ap_list(struct kvm_vcpu *vcpu) kvm_vcpu 337 virt/kvm/arm/vgic/vgic.c struct kvm_vcpu *vcpu; kvm_vcpu 440 virt/kvm/arm/vgic/vgic.c struct kvm_vcpu *vcpu; kvm_vcpu 480 virt/kvm/arm/vgic/vgic.c static int kvm_vgic_map_irq(struct kvm_vcpu *vcpu, struct vgic_irq *irq, kvm_vcpu 514 virt/kvm/arm/vgic/vgic.c int kvm_vgic_map_phys_irq(struct kvm_vcpu *vcpu, unsigned int host_irq, kvm_vcpu 540 virt/kvm/arm/vgic/vgic.c void kvm_vgic_reset_mapped_irq(struct kvm_vcpu *vcpu, u32 vintid) kvm_vcpu 557 virt/kvm/arm/vgic/vgic.c int kvm_vgic_unmap_phys_irq(struct kvm_vcpu *vcpu, unsigned int vintid) kvm_vcpu 586 virt/kvm/arm/vgic/vgic.c int kvm_vgic_set_owner(struct kvm_vcpu *vcpu, unsigned int intid, void *owner) kvm_vcpu 618 virt/kvm/arm/vgic/vgic.c static void vgic_prune_ap_list(struct kvm_vcpu *vcpu) kvm_vcpu 629 virt/kvm/arm/vgic/vgic.c struct kvm_vcpu *target_vcpu, *vcpuA, *vcpuB; kvm_vcpu 719 virt/kvm/arm/vgic/vgic.c static inline void vgic_fold_lr_state(struct kvm_vcpu *vcpu) kvm_vcpu 728 virt/kvm/arm/vgic/vgic.c static inline void vgic_populate_lr(struct kvm_vcpu *vcpu, kvm_vcpu 739 virt/kvm/arm/vgic/vgic.c static inline void vgic_clear_lr(struct kvm_vcpu *vcpu, int lr) kvm_vcpu 747 virt/kvm/arm/vgic/vgic.c static inline void vgic_set_underflow(struct kvm_vcpu *vcpu) kvm_vcpu 756 virt/kvm/arm/vgic/vgic.c static int compute_ap_list_depth(struct kvm_vcpu *vcpu, kvm_vcpu 782 virt/kvm/arm/vgic/vgic.c static void vgic_flush_lr_state(struct kvm_vcpu *vcpu) kvm_vcpu 847 virt/kvm/arm/vgic/vgic.c static inline void vgic_save_state(struct kvm_vcpu *vcpu) kvm_vcpu 856 virt/kvm/arm/vgic/vgic.c void kvm_vgic_sync_hwstate(struct kvm_vcpu *vcpu) kvm_vcpu 874 virt/kvm/arm/vgic/vgic.c static inline void vgic_restore_state(struct kvm_vcpu *vcpu) kvm_vcpu 883 virt/kvm/arm/vgic/vgic.c void kvm_vgic_flush_hwstate(struct kvm_vcpu *vcpu) kvm_vcpu 915 virt/kvm/arm/vgic/vgic.c void kvm_vgic_load(struct kvm_vcpu *vcpu) kvm_vcpu 926 virt/kvm/arm/vgic/vgic.c void kvm_vgic_put(struct kvm_vcpu *vcpu) kvm_vcpu 937 virt/kvm/arm/vgic/vgic.c void kvm_vgic_vmcr_sync(struct kvm_vcpu *vcpu) kvm_vcpu 948 virt/kvm/arm/vgic/vgic.c int kvm_vgic_vcpu_pending_irq(struct kvm_vcpu *vcpu) kvm_vcpu 984 virt/kvm/arm/vgic/vgic.c struct kvm_vcpu *vcpu; kvm_vcpu 999 virt/kvm/arm/vgic/vgic.c bool kvm_vgic_map_is_active(struct kvm_vcpu *vcpu, unsigned int vintid) kvm_vcpu 151 virt/kvm/arm/vgic/vgic.h struct kvm_vcpu *vcpu; kvm_vcpu 160 virt/kvm/arm/vgic/vgic.h vgic_get_mmio_region(struct kvm_vcpu *vcpu, struct vgic_io_device *iodev, kvm_vcpu 162 virt/kvm/arm/vgic/vgic.h struct vgic_irq *vgic_get_irq(struct kvm *kvm, struct kvm_vcpu *vcpu, kvm_vcpu 176 virt/kvm/arm/vgic/vgic.h void vgic_v2_fold_lr_state(struct kvm_vcpu *vcpu); kvm_vcpu 177 virt/kvm/arm/vgic/vgic.h void vgic_v2_populate_lr(struct kvm_vcpu *vcpu, struct vgic_irq *irq, int lr); kvm_vcpu 178 virt/kvm/arm/vgic/vgic.h void vgic_v2_clear_lr(struct kvm_vcpu *vcpu, int lr); kvm_vcpu 179 virt/kvm/arm/vgic/vgic.h void vgic_v2_set_underflow(struct kvm_vcpu *vcpu); kvm_vcpu 180 virt/kvm/arm/vgic/vgic.h void vgic_v2_set_npie(struct kvm_vcpu *vcpu); kvm_vcpu 182 virt/kvm/arm/vgic/vgic.h int vgic_v2_dist_uaccess(struct kvm_vcpu *vcpu, bool is_write, kvm_vcpu 184 virt/kvm/arm/vgic/vgic.h int vgic_v2_cpuif_uaccess(struct kvm_vcpu *vcpu, bool is_write, kvm_vcpu 186 virt/kvm/arm/vgic/vgic.h void vgic_v2_set_vmcr(struct kvm_vcpu *vcpu, struct vgic_vmcr *vmcr); kvm_vcpu 187 virt/kvm/arm/vgic/vgic.h void vgic_v2_get_vmcr(struct kvm_vcpu *vcpu, struct vgic_vmcr *vmcr); kvm_vcpu 188 virt/kvm/arm/vgic/vgic.h void vgic_v2_enable(struct kvm_vcpu *vcpu); kvm_vcpu 195 virt/kvm/arm/vgic/vgic.h void vgic_v2_load(struct kvm_vcpu *vcpu); kvm_vcpu 196 virt/kvm/arm/vgic/vgic.h void vgic_v2_put(struct kvm_vcpu *vcpu); kvm_vcpu 197 virt/kvm/arm/vgic/vgic.h void vgic_v2_vmcr_sync(struct kvm_vcpu *vcpu); kvm_vcpu 199 virt/kvm/arm/vgic/vgic.h void vgic_v2_save_state(struct kvm_vcpu *vcpu); kvm_vcpu 200 virt/kvm/arm/vgic/vgic.h void vgic_v2_restore_state(struct kvm_vcpu *vcpu); kvm_vcpu 210 virt/kvm/arm/vgic/vgic.h void vgic_v3_fold_lr_state(struct kvm_vcpu *vcpu); kvm_vcpu 211 virt/kvm/arm/vgic/vgic.h void vgic_v3_populate_lr(struct kvm_vcpu *vcpu, struct vgic_irq *irq, int lr); kvm_vcpu 212 virt/kvm/arm/vgic/vgic.h void vgic_v3_clear_lr(struct kvm_vcpu *vcpu, int lr); kvm_vcpu 213 virt/kvm/arm/vgic/vgic.h void vgic_v3_set_underflow(struct kvm_vcpu *vcpu); kvm_vcpu 214 virt/kvm/arm/vgic/vgic.h void vgic_v3_set_npie(struct kvm_vcpu *vcpu); kvm_vcpu 215 virt/kvm/arm/vgic/vgic.h void vgic_v3_set_vmcr(struct kvm_vcpu *vcpu, struct vgic_vmcr *vmcr); kvm_vcpu 216 virt/kvm/arm/vgic/vgic.h void vgic_v3_get_vmcr(struct kvm_vcpu *vcpu, struct vgic_vmcr *vmcr); kvm_vcpu 217 virt/kvm/arm/vgic/vgic.h void vgic_v3_enable(struct kvm_vcpu *vcpu); kvm_vcpu 223 virt/kvm/arm/vgic/vgic.h int vgic_register_redist_iodev(struct kvm_vcpu *vcpu); kvm_vcpu 226 virt/kvm/arm/vgic/vgic.h void vgic_v3_load(struct kvm_vcpu *vcpu); kvm_vcpu 227 virt/kvm/arm/vgic/vgic.h void vgic_v3_put(struct kvm_vcpu *vcpu); kvm_vcpu 228 virt/kvm/arm/vgic/vgic.h void vgic_v3_vmcr_sync(struct kvm_vcpu *vcpu); kvm_vcpu 232 virt/kvm/arm/vgic/vgic.h void vgic_enable_lpis(struct kvm_vcpu *vcpu); kvm_vcpu 233 virt/kvm/arm/vgic/vgic.h void vgic_flush_pending_lpis(struct kvm_vcpu *vcpu); kvm_vcpu 236 virt/kvm/arm/vgic/vgic.h int vgic_v3_dist_uaccess(struct kvm_vcpu *vcpu, bool is_write, kvm_vcpu 238 virt/kvm/arm/vgic/vgic.h int vgic_v3_redist_uaccess(struct kvm_vcpu *vcpu, bool is_write, kvm_vcpu 240 virt/kvm/arm/vgic/vgic.h int vgic_v3_cpu_sysregs_uaccess(struct kvm_vcpu *vcpu, bool is_write, kvm_vcpu 242 virt/kvm/arm/vgic/vgic.h int vgic_v3_has_cpu_sysregs_attr(struct kvm_vcpu *vcpu, bool is_write, u64 id, kvm_vcpu 244 virt/kvm/arm/vgic/vgic.h int vgic_v3_line_level_info_uaccess(struct kvm_vcpu *vcpu, bool is_write, kvm_vcpu 247 virt/kvm/arm/vgic/vgic.h void vgic_set_vmcr(struct kvm_vcpu *vcpu, struct vgic_vmcr *vmcr); kvm_vcpu 248 virt/kvm/arm/vgic/vgic.h void vgic_get_vmcr(struct kvm_vcpu *vcpu, struct vgic_vmcr *vmcr); kvm_vcpu 258 virt/kvm/arm/vgic/vgic.h static inline int vgic_v3_max_apr_idx(struct kvm_vcpu *vcpu) kvm_vcpu 307 virt/kvm/arm/vgic/vgic.h int vgic_copy_lpi_list(struct kvm *kvm, struct kvm_vcpu *vcpu, u32 **intid_ptr); kvm_vcpu 319 virt/kvm/arm/vgic/vgic.h int vgic_v4_sync_hwstate(struct kvm_vcpu *vcpu); kvm_vcpu 320 virt/kvm/arm/vgic/vgic.h int vgic_v4_flush_hwstate(struct kvm_vcpu *vcpu); kvm_vcpu 20 virt/kvm/async_pf.c static inline void kvm_async_page_present_sync(struct kvm_vcpu *vcpu, kvm_vcpu 27 virt/kvm/async_pf.c static inline void kvm_async_page_present_async(struct kvm_vcpu *vcpu, kvm_vcpu 53 virt/kvm/async_pf.c void kvm_async_pf_vcpu_init(struct kvm_vcpu *vcpu) kvm_vcpu 65 virt/kvm/async_pf.c struct kvm_vcpu *vcpu = apf->vcpu; kvm_vcpu 104 virt/kvm/async_pf.c void kvm_clear_async_pf_completion_queue(struct kvm_vcpu *vcpu) kvm_vcpu 147 virt/kvm/async_pf.c void kvm_check_async_pf_completion(struct kvm_vcpu *vcpu) kvm_vcpu 168 virt/kvm/async_pf.c int kvm_setup_async_pf(struct kvm_vcpu *vcpu, gpa_t cr2_or_gpa, kvm_vcpu 215 virt/kvm/async_pf.c int kvm_async_pf_wakeup_all(struct kvm_vcpu *vcpu) kvm_vcpu 17 virt/kvm/async_pf.h void kvm_async_pf_vcpu_init(struct kvm_vcpu *vcpu); kvm_vcpu 64 virt/kvm/coalesced_mmio.c static int coalesced_mmio_write(struct kvm_vcpu *vcpu, kvm_vcpu 729 virt/kvm/eventfd.c ioeventfd_write(struct kvm_vcpu *vcpu, struct kvm_io_device *this, gpa_t addr, kvm_vcpu 196 virt/kvm/kvm_main.c void vcpu_load(struct kvm_vcpu *vcpu) kvm_vcpu 205 virt/kvm/kvm_main.c void vcpu_put(struct kvm_vcpu *vcpu) kvm_vcpu 215 virt/kvm/kvm_main.c static bool kvm_request_needs_ipi(struct kvm_vcpu *vcpu, unsigned req) kvm_vcpu 252 virt/kvm/kvm_main.c struct kvm_vcpu *vcpu; kvm_vcpu 324 virt/kvm/kvm_main.c int kvm_vcpu_init(struct kvm_vcpu *vcpu, struct kvm *kvm, unsigned id) kvm_vcpu 364 virt/kvm/kvm_main.c void kvm_vcpu_uninit(struct kvm_vcpu *vcpu) kvm_vcpu 1386 virt/kvm/kvm_main.c struct kvm_memory_slot *kvm_vcpu_gfn_to_memslot(struct kvm_vcpu *vcpu, gfn_t gfn) kvm_vcpu 1403 virt/kvm/kvm_main.c unsigned long kvm_host_page_size(struct kvm_vcpu *vcpu, gfn_t gfn) kvm_vcpu 1466 virt/kvm/kvm_main.c unsigned long kvm_vcpu_gfn_to_hva(struct kvm_vcpu *vcpu, gfn_t gfn) kvm_vcpu 1498 virt/kvm/kvm_main.c unsigned long kvm_vcpu_gfn_to_hva_prot(struct kvm_vcpu *vcpu, gfn_t gfn, bool *writable) kvm_vcpu 1760 virt/kvm/kvm_main.c kvm_pfn_t kvm_vcpu_gfn_to_pfn_atomic(struct kvm_vcpu *vcpu, gfn_t gfn) kvm_vcpu 1772 virt/kvm/kvm_main.c kvm_pfn_t kvm_vcpu_gfn_to_pfn(struct kvm_vcpu *vcpu, gfn_t gfn) kvm_vcpu 1898 virt/kvm/kvm_main.c int kvm_map_gfn(struct kvm_vcpu *vcpu, gfn_t gfn, struct kvm_host_map *map, kvm_vcpu 1906 virt/kvm/kvm_main.c int kvm_vcpu_map(struct kvm_vcpu *vcpu, gfn_t gfn, struct kvm_host_map *map) kvm_vcpu 1949 virt/kvm/kvm_main.c int kvm_unmap_gfn(struct kvm_vcpu *vcpu, struct kvm_host_map *map, kvm_vcpu 1958 virt/kvm/kvm_main.c void kvm_vcpu_unmap(struct kvm_vcpu *vcpu, struct kvm_host_map *map, bool dirty) kvm_vcpu 1965 virt/kvm/kvm_main.c struct page *kvm_vcpu_gfn_to_page(struct kvm_vcpu *vcpu, gfn_t gfn) kvm_vcpu 2061 virt/kvm/kvm_main.c int kvm_vcpu_read_guest_page(struct kvm_vcpu *vcpu, gfn_t gfn, void *data, kvm_vcpu 2090 virt/kvm/kvm_main.c int kvm_vcpu_read_guest(struct kvm_vcpu *vcpu, gpa_t gpa, void *data, unsigned long len) kvm_vcpu 2138 virt/kvm/kvm_main.c int kvm_vcpu_read_guest_atomic(struct kvm_vcpu *vcpu, gpa_t gpa, kvm_vcpu 2174 virt/kvm/kvm_main.c int kvm_vcpu_write_guest_page(struct kvm_vcpu *vcpu, gfn_t gfn, kvm_vcpu 2204 virt/kvm/kvm_main.c int kvm_vcpu_write_guest(struct kvm_vcpu *vcpu, gpa_t gpa, const void *data, kvm_vcpu 2377 virt/kvm/kvm_main.c void kvm_vcpu_mark_page_dirty(struct kvm_vcpu *vcpu, gfn_t gfn) kvm_vcpu 2386 virt/kvm/kvm_main.c void kvm_sigset_activate(struct kvm_vcpu *vcpu) kvm_vcpu 2400 virt/kvm/kvm_main.c void kvm_sigset_deactivate(struct kvm_vcpu *vcpu) kvm_vcpu 2409 virt/kvm/kvm_main.c static void grow_halt_poll_ns(struct kvm_vcpu *vcpu) kvm_vcpu 2431 virt/kvm/kvm_main.c static void shrink_halt_poll_ns(struct kvm_vcpu *vcpu) kvm_vcpu 2446 virt/kvm/kvm_main.c static int kvm_vcpu_check_block(struct kvm_vcpu *vcpu) kvm_vcpu 2469 virt/kvm/kvm_main.c void kvm_vcpu_block(struct kvm_vcpu *vcpu) kvm_vcpu 2537 virt/kvm/kvm_main.c bool kvm_vcpu_wake_up(struct kvm_vcpu *vcpu) kvm_vcpu 2557 virt/kvm/kvm_main.c void kvm_vcpu_kick(struct kvm_vcpu *vcpu) kvm_vcpu 2574 virt/kvm/kvm_main.c int kvm_vcpu_yield_to(struct kvm_vcpu *target) kvm_vcpu 2616 virt/kvm/kvm_main.c static bool kvm_vcpu_eligible_for_directed_yield(struct kvm_vcpu *vcpu) kvm_vcpu 2638 virt/kvm/kvm_main.c bool __weak kvm_arch_dy_runnable(struct kvm_vcpu *vcpu) kvm_vcpu 2643 virt/kvm/kvm_main.c static bool vcpu_dy_runnable(struct kvm_vcpu *vcpu) kvm_vcpu 2656 virt/kvm/kvm_main.c void kvm_vcpu_on_spin(struct kvm_vcpu *me, bool yield_to_kernel_mode) kvm_vcpu 2659 virt/kvm/kvm_main.c struct kvm_vcpu *vcpu; kvm_vcpu 2713 virt/kvm/kvm_main.c struct kvm_vcpu *vcpu = vmf->vma->vm_file->private_data; kvm_vcpu 2745 virt/kvm/kvm_main.c struct kvm_vcpu *vcpu = filp->private_data; kvm_vcpu 2763 virt/kvm/kvm_main.c static int create_vcpu_fd(struct kvm_vcpu *vcpu) kvm_vcpu 2771 virt/kvm/kvm_main.c static void kvm_create_vcpu_debugfs(struct kvm_vcpu *vcpu) kvm_vcpu 2793 virt/kvm/kvm_main.c struct kvm_vcpu *vcpu; kvm_vcpu 2862 virt/kvm/kvm_main.c static int kvm_vcpu_ioctl_set_sigmask(struct kvm_vcpu *vcpu, sigset_t *sigset) kvm_vcpu 2876 virt/kvm/kvm_main.c struct kvm_vcpu *vcpu = filp->private_data; kvm_vcpu 3085 virt/kvm/kvm_main.c struct kvm_vcpu *vcpu = filp->private_data; kvm_vcpu 3854 virt/kvm/kvm_main.c static int __kvm_io_bus_write(struct kvm_vcpu *vcpu, struct kvm_io_bus *bus, kvm_vcpu 3875 virt/kvm/kvm_main.c int kvm_io_bus_write(struct kvm_vcpu *vcpu, enum kvm_bus bus_idx, gpa_t addr, kvm_vcpu 3896 virt/kvm/kvm_main.c int kvm_io_bus_write_cookie(struct kvm_vcpu *vcpu, enum kvm_bus bus_idx, kvm_vcpu 3925 virt/kvm/kvm_main.c static int __kvm_io_bus_read(struct kvm_vcpu *vcpu, struct kvm_io_bus *bus, kvm_vcpu 3946 virt/kvm/kvm_main.c int kvm_io_bus_read(struct kvm_vcpu *vcpu, enum kvm_bus bus_idx, gpa_t addr, kvm_vcpu 4149 virt/kvm/kvm_main.c struct kvm_vcpu *vcpu; kvm_vcpu 4163 virt/kvm/kvm_main.c struct kvm_vcpu *vcpu; kvm_vcpu 4367 virt/kvm/kvm_main.c struct kvm_vcpu *preempt_notifier_to_vcpu(struct preempt_notifier *pn) kvm_vcpu 4369 virt/kvm/kvm_main.c return container_of(pn, struct kvm_vcpu, preempt_notifier); kvm_vcpu 4374 virt/kvm/kvm_main.c struct kvm_vcpu *vcpu = preempt_notifier_to_vcpu(pn); kvm_vcpu 4387 virt/kvm/kvm_main.c struct kvm_vcpu *vcpu = preempt_notifier_to_vcpu(pn); kvm_vcpu 4445 virt/kvm/kvm_main.c vcpu_align = __alignof__(struct kvm_vcpu); kvm_vcpu 4449 virt/kvm/kvm_main.c offsetof(struct kvm_vcpu, arch), kvm_vcpu 4450 virt/kvm/kvm_main.c sizeof_field(struct kvm_vcpu, arch),