run 19 arch/arm/include/asm/kvm_coproc.h int kvm_handle_cp10_id(struct kvm_vcpu *vcpu, struct kvm_run *run); run 20 arch/arm/include/asm/kvm_coproc.h int kvm_handle_cp_0_13_access(struct kvm_vcpu *vcpu, struct kvm_run *run); run 21 arch/arm/include/asm/kvm_coproc.h int kvm_handle_cp14_load_store(struct kvm_vcpu *vcpu, struct kvm_run *run); run 22 arch/arm/include/asm/kvm_coproc.h int kvm_handle_cp14_32(struct kvm_vcpu *vcpu, struct kvm_run *run); run 23 arch/arm/include/asm/kvm_coproc.h int kvm_handle_cp14_64(struct kvm_vcpu *vcpu, struct kvm_run *run); run 24 arch/arm/include/asm/kvm_coproc.h int kvm_handle_cp15_32(struct kvm_vcpu *vcpu, struct kvm_run *run); run 25 arch/arm/include/asm/kvm_coproc.h int kvm_handle_cp15_64(struct kvm_vcpu *vcpu, struct kvm_run *run); run 287 arch/arm/include/asm/kvm_host.h int handle_exit(struct kvm_vcpu *vcpu, struct kvm_run *run, run 290 arch/arm/include/asm/kvm_host.h static inline void handle_exit_early(struct kvm_vcpu *vcpu, struct kvm_run *run, run 24 arch/arm/include/asm/kvm_mmio.h int kvm_handle_mmio_return(struct kvm_vcpu *vcpu, struct kvm_run *run); run 25 arch/arm/include/asm/kvm_mmio.h int io_mem_abort(struct kvm_vcpu *vcpu, struct kvm_run *run, run 60 arch/arm/include/asm/kvm_mmu.h int kvm_handle_guest_abort(struct kvm_vcpu *vcpu, struct kvm_run *run); run 81 arch/arm/kvm/coproc.c int kvm_handle_cp10_id(struct kvm_vcpu *vcpu, struct kvm_run *run) run 87 arch/arm/kvm/coproc.c int kvm_handle_cp_0_13_access(struct kvm_vcpu *vcpu, struct kvm_run *run) run 97 arch/arm/kvm/coproc.c int kvm_handle_cp14_load_store(struct kvm_vcpu *vcpu, struct kvm_run *run) run 629 arch/arm/kvm/coproc.c int kvm_handle_cp15_64(struct kvm_vcpu *vcpu, struct kvm_run *run) run 641 arch/arm/kvm/coproc.c int kvm_handle_cp14_64(struct kvm_vcpu *vcpu, struct kvm_run *run) run 694 arch/arm/kvm/coproc.c int kvm_handle_cp15_32(struct kvm_vcpu *vcpu, struct kvm_run *run) run 705 arch/arm/kvm/coproc.c int kvm_handle_cp14_32(struct kvm_vcpu *vcpu, struct kvm_run *run) run 19 arch/arm/kvm/handle_exit.c static int handle_hvc(struct kvm_vcpu *vcpu, struct kvm_run *run) run 36 arch/arm/kvm/handle_exit.c static int handle_smc(struct kvm_vcpu *vcpu, struct kvm_run *run) run 62 arch/arm/kvm/handle_exit.c static int kvm_handle_wfx(struct kvm_vcpu *vcpu, struct kvm_run *run) run 80 arch/arm/kvm/handle_exit.c static int kvm_handle_unknown_ec(struct kvm_vcpu *vcpu, struct kvm_run *run) run 118 arch/arm/kvm/handle_exit.c int handle_exit(struct kvm_vcpu *vcpu, struct kvm_run *run, run 157 arch/arm/kvm/handle_exit.c return exit_handler(vcpu, run); run 167 arch/arm/kvm/handle_exit.c run->exit_reason = KVM_EXIT_FAIL_ENTRY; run 172 arch/arm/kvm/handle_exit.c run->exit_reason = KVM_EXIT_INTERNAL_ERROR; run 30 arch/arm64/include/asm/kvm_coproc.h int kvm_handle_cp14_load_store(struct kvm_vcpu *vcpu, struct kvm_run *run); run 31 arch/arm64/include/asm/kvm_coproc.h int kvm_handle_cp14_32(struct kvm_vcpu *vcpu, struct kvm_run *run); run 32 arch/arm64/include/asm/kvm_coproc.h int kvm_handle_cp14_64(struct kvm_vcpu *vcpu, struct kvm_run *run); run 33 arch/arm64/include/asm/kvm_coproc.h int kvm_handle_cp15_32(struct kvm_vcpu *vcpu, struct kvm_run *run); run 34 arch/arm64/include/asm/kvm_coproc.h int kvm_handle_cp15_64(struct kvm_vcpu *vcpu, struct kvm_run *run); run 35 arch/arm64/include/asm/kvm_coproc.h int kvm_handle_sys_reg(struct kvm_vcpu *vcpu, struct kvm_run *run); run 475 arch/arm64/include/asm/kvm_host.h int handle_exit(struct kvm_vcpu *vcpu, struct kvm_run *run, run 477 arch/arm64/include/asm/kvm_host.h void handle_exit_early(struct kvm_vcpu *vcpu, struct kvm_run *run, run 23 arch/arm64/include/asm/kvm_mmio.h int kvm_handle_mmio_return(struct kvm_vcpu *vcpu, struct kvm_run *run); run 24 arch/arm64/include/asm/kvm_mmio.h int io_mem_abort(struct kvm_vcpu *vcpu, struct kvm_run *run, run 161 arch/arm64/include/asm/kvm_mmu.h int kvm_handle_guest_abort(struct kvm_vcpu *vcpu, struct kvm_run *run); run 36 arch/arm64/kvm/handle_exit.c static int handle_hvc(struct kvm_vcpu *vcpu, struct kvm_run *run) run 53 arch/arm64/kvm/handle_exit.c static int handle_smc(struct kvm_vcpu *vcpu, struct kvm_run *run) run 72 arch/arm64/kvm/handle_exit.c static int handle_no_fpsimd(struct kvm_vcpu *vcpu, struct kvm_run *run) run 90 arch/arm64/kvm/handle_exit.c static int kvm_handle_wfx(struct kvm_vcpu *vcpu, struct kvm_run *run) run 120 arch/arm64/kvm/handle_exit.c static int kvm_handle_guest_debug(struct kvm_vcpu *vcpu, struct kvm_run *run) run 125 arch/arm64/kvm/handle_exit.c run->exit_reason = KVM_EXIT_DEBUG; run 126 arch/arm64/kvm/handle_exit.c run->debug.arch.hsr = hsr; run 130 arch/arm64/kvm/handle_exit.c run->debug.arch.far = vcpu->arch.fault.far_el2; run 147 arch/arm64/kvm/handle_exit.c static int kvm_handle_unknown_ec(struct kvm_vcpu *vcpu, struct kvm_run *run) run 158 arch/arm64/kvm/handle_exit.c static int handle_sve(struct kvm_vcpu *vcpu, struct kvm_run *run) run 181 arch/arm64/kvm/handle_exit.c static int kvm_handle_ptrauth(struct kvm_vcpu *vcpu, struct kvm_run *run) run 226 arch/arm64/kvm/handle_exit.c static int handle_trap_exceptions(struct kvm_vcpu *vcpu, struct kvm_run *run) run 241 arch/arm64/kvm/handle_exit.c handled = exit_handler(vcpu, run); run 251 arch/arm64/kvm/handle_exit.c int handle_exit(struct kvm_vcpu *vcpu, struct kvm_run *run, run 279 arch/arm64/kvm/handle_exit.c return handle_trap_exceptions(vcpu, run); run 285 arch/arm64/kvm/handle_exit.c run->exit_reason = KVM_EXIT_FAIL_ENTRY; run 292 arch/arm64/kvm/handle_exit.c run->exit_reason = KVM_EXIT_FAIL_ENTRY; run 297 arch/arm64/kvm/handle_exit.c run->exit_reason = KVM_EXIT_INTERNAL_ERROR; run 303 arch/arm64/kvm/handle_exit.c void handle_exit_early(struct kvm_vcpu *vcpu, struct kvm_run *run, run 2027 arch/arm64/kvm/sys_regs.c int kvm_handle_cp14_load_store(struct kvm_vcpu *vcpu, struct kvm_run *run) run 2206 arch/arm64/kvm/sys_regs.c int kvm_handle_cp15_64(struct kvm_vcpu *vcpu, struct kvm_run *run) run 2217 arch/arm64/kvm/sys_regs.c int kvm_handle_cp15_32(struct kvm_vcpu *vcpu, struct kvm_run *run) run 2228 arch/arm64/kvm/sys_regs.c int kvm_handle_cp14_64(struct kvm_vcpu *vcpu, struct kvm_run *run) run 2235 arch/arm64/kvm/sys_regs.c int kvm_handle_cp14_32(struct kvm_vcpu *vcpu, struct kvm_run *run) run 2287 arch/arm64/kvm/sys_regs.c int kvm_handle_sys_reg(struct kvm_vcpu *vcpu, struct kvm_run *run) run 317 arch/mips/include/asm/kvm_host.h int (*vcpu_run)(struct kvm_run *run, struct kvm_vcpu *vcpu); run 819 arch/mips/include/asm/kvm_host.h int (*vcpu_run)(struct kvm_run *run, struct kvm_vcpu *vcpu); run 820 arch/mips/include/asm/kvm_host.h void (*vcpu_reenter)(struct kvm_run *run, struct kvm_vcpu *vcpu); run 828 arch/mips/include/asm/kvm_host.h extern int kvm_mips_handle_exit(struct kvm_run *run, struct kvm_vcpu *vcpu); run 875 arch/mips/include/asm/kvm_host.h struct kvm_run *run, run 982 arch/mips/include/asm/kvm_host.h struct kvm_run *run, run 989 arch/mips/include/asm/kvm_host.h struct kvm_run *run, run 994 arch/mips/include/asm/kvm_host.h struct kvm_run *run, run 999 arch/mips/include/asm/kvm_host.h struct kvm_run *run, run 1004 arch/mips/include/asm/kvm_host.h struct kvm_run *run, run 1009 arch/mips/include/asm/kvm_host.h struct kvm_run *run, run 1014 arch/mips/include/asm/kvm_host.h struct kvm_run *run, run 1019 arch/mips/include/asm/kvm_host.h struct kvm_run *run, run 1024 arch/mips/include/asm/kvm_host.h struct kvm_run *run, run 1029 arch/mips/include/asm/kvm_host.h struct kvm_run *run, run 1034 arch/mips/include/asm/kvm_host.h struct kvm_run *run, run 1039 arch/mips/include/asm/kvm_host.h struct kvm_run *run, run 1044 arch/mips/include/asm/kvm_host.h struct kvm_run *run, run 1049 arch/mips/include/asm/kvm_host.h struct kvm_run *run, run 1054 arch/mips/include/asm/kvm_host.h struct kvm_run *run, run 1058 arch/mips/include/asm/kvm_host.h struct kvm_run *run); run 1087 arch/mips/include/asm/kvm_host.h struct kvm_run *run, run 1093 arch/mips/include/asm/kvm_host.h struct kvm_run *run, run 1098 arch/mips/include/asm/kvm_host.h struct kvm_run *run, run 1102 arch/mips/include/asm/kvm_host.h struct kvm_run *run, run 1106 arch/mips/include/asm/kvm_host.h struct kvm_run *run, run 987 arch/mips/kvm/emulate.c vcpu->run->exit_reason = KVM_EXIT_IRQ_WINDOW_OPEN; run 1265 arch/mips/kvm/emulate.c struct kvm_run *run, run 1600 arch/mips/kvm/emulate.c struct kvm_run *run, run 1605 arch/mips/kvm/emulate.c void *data = run->mmio.data; run 1619 arch/mips/kvm/emulate.c run->mmio.phys_addr = kvm_mips_callbacks->gva_to_gpa( run 1621 arch/mips/kvm/emulate.c if (run->mmio.phys_addr == KVM_INVALID_ADDR) run 1627 arch/mips/kvm/emulate.c run->mmio.len = 8; run 1637 arch/mips/kvm/emulate.c run->mmio.len = 4; run 1646 arch/mips/kvm/emulate.c run->mmio.len = 2; run 1655 arch/mips/kvm/emulate.c run->mmio.len = 1; run 1669 arch/mips/kvm/emulate.c run->mmio.is_write = 1; run 1681 arch/mips/kvm/emulate.c u32 cause, struct kvm_run *run, run 1705 arch/mips/kvm/emulate.c run->mmio.phys_addr = kvm_mips_callbacks->gva_to_gpa( run 1707 arch/mips/kvm/emulate.c if (run->mmio.phys_addr == KVM_INVALID_ADDR) run 1714 arch/mips/kvm/emulate.c run->mmio.len = 8; run 1722 arch/mips/kvm/emulate.c run->mmio.len = 4; run 1729 arch/mips/kvm/emulate.c run->mmio.len = 2; run 1736 arch/mips/kvm/emulate.c run->mmio.len = 1; run 1746 arch/mips/kvm/emulate.c run->mmio.is_write = 0; run 1755 arch/mips/kvm/emulate.c struct kvm_run *run, run 1783 arch/mips/kvm/emulate.c kvm_mips_emulate_tlbmiss_ld(cause, NULL, run, vcpu); run 1789 arch/mips/kvm/emulate.c kvm_mips_emulate_tlbinv_ld(cause, NULL, run, vcpu); run 1799 arch/mips/kvm/emulate.c struct kvm_run *run, run 1889 arch/mips/kvm/emulate.c curr_pc, va, run, vcpu, cause); run 1902 arch/mips/kvm/emulate.c curr_pc, va, run, vcpu, cause); run 1906 arch/mips/kvm/emulate.c curr_pc, va, run, vcpu, cause); run 1932 arch/mips/kvm/emulate.c struct kvm_run *run, run 1948 arch/mips/kvm/emulate.c er = kvm_mips_emulate_CP0(inst, opc, cause, run, vcpu); run 1955 arch/mips/kvm/emulate.c er = kvm_mips_emulate_cache(inst, opc, cause, run, vcpu); run 1963 arch/mips/kvm/emulate.c er = kvm_mips_emulate_cache(inst, opc, cause, run, run 2003 arch/mips/kvm/emulate.c struct kvm_run *run, run 2038 arch/mips/kvm/emulate.c struct kvm_run *run, run 2082 arch/mips/kvm/emulate.c struct kvm_run *run, run 2124 arch/mips/kvm/emulate.c struct kvm_run *run, run 2166 arch/mips/kvm/emulate.c struct kvm_run *run, run 2207 arch/mips/kvm/emulate.c struct kvm_run *run, run 2247 arch/mips/kvm/emulate.c struct kvm_run *run, run 2276 arch/mips/kvm/emulate.c struct kvm_run *run, run 2311 arch/mips/kvm/emulate.c struct kvm_run *run, run 2346 arch/mips/kvm/emulate.c struct kvm_run *run, run 2381 arch/mips/kvm/emulate.c struct kvm_run *run, run 2416 arch/mips/kvm/emulate.c struct kvm_run *run, run 2451 arch/mips/kvm/emulate.c struct kvm_run *run, run 2485 arch/mips/kvm/emulate.c struct kvm_run *run, run 2574 arch/mips/kvm/emulate.c return kvm_mips_emulate_ri_exc(cause, opc, run, vcpu); run 2578 arch/mips/kvm/emulate.c struct kvm_run *run) run 2583 arch/mips/kvm/emulate.c if (run->mmio.len > sizeof(*gpr)) { run 2584 arch/mips/kvm/emulate.c kvm_err("Bad MMIO length: %d", run->mmio.len); run 2592 arch/mips/kvm/emulate.c switch (run->mmio.len) { run 2594 arch/mips/kvm/emulate.c *gpr = *(s64 *)run->mmio.data; run 2599 arch/mips/kvm/emulate.c *gpr = *(s32 *)run->mmio.data; run 2601 arch/mips/kvm/emulate.c *gpr = *(u32 *)run->mmio.data; run 2606 arch/mips/kvm/emulate.c *gpr = *(s16 *) run->mmio.data; run 2608 arch/mips/kvm/emulate.c *gpr = *(u16 *)run->mmio.data; run 2613 arch/mips/kvm/emulate.c *gpr = *(s8 *) run->mmio.data; run 2615 arch/mips/kvm/emulate.c *gpr = *(u8 *) run->mmio.data; run 2625 arch/mips/kvm/emulate.c struct kvm_run *run, run 2663 arch/mips/kvm/emulate.c struct kvm_run *run, run 2745 arch/mips/kvm/emulate.c kvm_mips_emulate_exc(cause, opc, run, vcpu); run 2759 arch/mips/kvm/emulate.c struct kvm_run *run, run 2783 arch/mips/kvm/emulate.c er = kvm_mips_emulate_tlbmiss_ld(cause, opc, run, vcpu); run 2785 arch/mips/kvm/emulate.c er = kvm_mips_emulate_tlbmiss_st(cause, opc, run, vcpu); run 2800 arch/mips/kvm/emulate.c er = kvm_mips_emulate_tlbinv_ld(cause, opc, run, run 2803 arch/mips/kvm/emulate.c er = kvm_mips_emulate_tlbinv_st(cause, opc, run, run 634 arch/mips/kvm/entry.c UASM_i_LW(&p, S0, offsetof(struct kvm_vcpu, run), S1); run 435 arch/mips/kvm/mips.c int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu, struct kvm_run *run) run 445 arch/mips/kvm/mips.c kvm_mips_complete_mmio_load(vcpu, run); run 449 arch/mips/kvm/mips.c if (run->immediate_exit) run 466 arch/mips/kvm/mips.c r = kvm_mips_callbacks->vcpu_run(run, vcpu); run 1281 arch/mips/kvm/mips.c int kvm_mips_handle_exit(struct kvm_run *run, struct kvm_vcpu *vcpu) run 1298 arch/mips/kvm/mips.c run->exit_reason = KVM_EXIT_UNKNOWN; run 1299 arch/mips/kvm/mips.c run->ready_for_interrupt_injection = 1; run 1310 arch/mips/kvm/mips.c cause, opc, run, vcpu); run 1319 arch/mips/kvm/mips.c er = kvm_mips_check_privilege(cause, opc, run, vcpu); run 1323 arch/mips/kvm/mips.c run->exit_reason = KVM_EXIT_INTERNAL_ERROR; run 1347 arch/mips/kvm/mips.c if (run->exit_reason == KVM_EXIT_IRQ_WINDOW_OPEN) run 1432 arch/mips/kvm/mips.c run->exit_reason = KVM_EXIT_INTERNAL_ERROR; run 1450 arch/mips/kvm/mips.c run->exit_reason = KVM_EXIT_INTR; run 1468 arch/mips/kvm/mips.c kvm_mips_callbacks->vcpu_reenter(run, vcpu); run 63 arch/mips/kvm/trap_emul.c vcpu->run->exit_reason = KVM_EXIT_INTERNAL_ERROR; run 70 arch/mips/kvm/trap_emul.c struct kvm_run *run = vcpu->run; run 84 arch/mips/kvm/trap_emul.c er = kvm_mips_emulate_fpu_exc(cause, opc, run, vcpu); run 91 arch/mips/kvm/trap_emul.c er = kvm_mips_emulate_inst(cause, opc, run, vcpu); run 100 arch/mips/kvm/trap_emul.c run->exit_reason = KVM_EXIT_INTERNAL_ERROR; run 105 arch/mips/kvm/trap_emul.c run->exit_reason = KVM_EXIT_INTR; run 119 arch/mips/kvm/trap_emul.c static int kvm_mips_bad_load(u32 cause, u32 *opc, struct kvm_run *run, run 128 arch/mips/kvm/trap_emul.c run->exit_reason = KVM_EXIT_INTERNAL_ERROR; run 137 arch/mips/kvm/trap_emul.c run->exit_reason = KVM_EXIT_INTERNAL_ERROR; run 142 arch/mips/kvm/trap_emul.c er = kvm_mips_emulate_load(inst, cause, run, vcpu); run 145 arch/mips/kvm/trap_emul.c run->exit_reason = KVM_EXIT_INTERNAL_ERROR; run 147 arch/mips/kvm/trap_emul.c run->exit_reason = KVM_EXIT_MMIO; run 152 arch/mips/kvm/trap_emul.c static int kvm_mips_bad_store(u32 cause, u32 *opc, struct kvm_run *run, run 164 arch/mips/kvm/trap_emul.c run->exit_reason = KVM_EXIT_INTERNAL_ERROR; run 169 arch/mips/kvm/trap_emul.c er = kvm_mips_emulate_store(inst, cause, run, vcpu); run 172 arch/mips/kvm/trap_emul.c run->exit_reason = KVM_EXIT_INTERNAL_ERROR; run 174 arch/mips/kvm/trap_emul.c run->exit_reason = KVM_EXIT_MMIO; run 179 arch/mips/kvm/trap_emul.c static int kvm_mips_bad_access(u32 cause, u32 *opc, struct kvm_run *run, run 183 arch/mips/kvm/trap_emul.c return kvm_mips_bad_store(cause, opc, run, vcpu); run 185 arch/mips/kvm/trap_emul.c return kvm_mips_bad_load(cause, opc, run, vcpu); run 191 arch/mips/kvm/trap_emul.c struct kvm_run *run = vcpu->run; run 215 arch/mips/kvm/trap_emul.c run->exit_reason = KVM_EXIT_INTERNAL_ERROR; run 220 arch/mips/kvm/trap_emul.c run->exit_reason = KVM_EXIT_INTERNAL_ERROR; run 229 arch/mips/kvm/trap_emul.c kvm_mips_emulate_tlbmod(cause, opc, run, vcpu); run 236 arch/mips/kvm/trap_emul.c return kvm_mips_bad_store(cause, opc, run, vcpu); run 241 arch/mips/kvm/trap_emul.c return kvm_mips_bad_store(cause, opc, run, vcpu); run 245 arch/mips/kvm/trap_emul.c return kvm_mips_bad_store(cause, opc, run, vcpu); run 251 arch/mips/kvm/trap_emul.c struct kvm_run *run = vcpu->run; run 261 arch/mips/kvm/trap_emul.c run->exit_reason = KVM_EXIT_INTERNAL_ERROR; run 279 arch/mips/kvm/trap_emul.c er = kvm_mips_handle_tlbmiss(cause, opc, run, vcpu, store); run 283 arch/mips/kvm/trap_emul.c run->exit_reason = KVM_EXIT_INTERNAL_ERROR; run 292 arch/mips/kvm/trap_emul.c ret = kvm_mips_bad_access(cause, opc, run, vcpu, store); run 299 arch/mips/kvm/trap_emul.c ret = kvm_mips_bad_access(cause, opc, run, vcpu, store); run 305 arch/mips/kvm/trap_emul.c run->exit_reason = KVM_EXIT_INTERNAL_ERROR; run 323 arch/mips/kvm/trap_emul.c struct kvm_run *run = vcpu->run; run 331 arch/mips/kvm/trap_emul.c ret = kvm_mips_bad_store(cause, opc, run, vcpu); run 335 arch/mips/kvm/trap_emul.c run->exit_reason = KVM_EXIT_INTERNAL_ERROR; run 343 arch/mips/kvm/trap_emul.c struct kvm_run *run = vcpu->run; run 350 arch/mips/kvm/trap_emul.c ret = kvm_mips_bad_load(cause, opc, run, vcpu); run 354 arch/mips/kvm/trap_emul.c run->exit_reason = KVM_EXIT_INTERNAL_ERROR; run 362 arch/mips/kvm/trap_emul.c struct kvm_run *run = vcpu->run; run 368 arch/mips/kvm/trap_emul.c er = kvm_mips_emulate_syscall(cause, opc, run, vcpu); run 372 arch/mips/kvm/trap_emul.c run->exit_reason = KVM_EXIT_INTERNAL_ERROR; run 380 arch/mips/kvm/trap_emul.c struct kvm_run *run = vcpu->run; run 386 arch/mips/kvm/trap_emul.c er = kvm_mips_handle_ri(cause, opc, run, vcpu); run 390 arch/mips/kvm/trap_emul.c run->exit_reason = KVM_EXIT_INTERNAL_ERROR; run 398 arch/mips/kvm/trap_emul.c struct kvm_run *run = vcpu->run; run 404 arch/mips/kvm/trap_emul.c er = kvm_mips_emulate_bp_exc(cause, opc, run, vcpu); run 408 arch/mips/kvm/trap_emul.c run->exit_reason = KVM_EXIT_INTERNAL_ERROR; run 416 arch/mips/kvm/trap_emul.c struct kvm_run *run = vcpu->run; run 422 arch/mips/kvm/trap_emul.c er = kvm_mips_emulate_trap_exc(cause, opc, run, vcpu); run 426 arch/mips/kvm/trap_emul.c run->exit_reason = KVM_EXIT_INTERNAL_ERROR; run 434 arch/mips/kvm/trap_emul.c struct kvm_run *run = vcpu->run; run 440 arch/mips/kvm/trap_emul.c er = kvm_mips_emulate_msafpe_exc(cause, opc, run, vcpu); run 444 arch/mips/kvm/trap_emul.c run->exit_reason = KVM_EXIT_INTERNAL_ERROR; run 452 arch/mips/kvm/trap_emul.c struct kvm_run *run = vcpu->run; run 458 arch/mips/kvm/trap_emul.c er = kvm_mips_emulate_fpe_exc(cause, opc, run, vcpu); run 462 arch/mips/kvm/trap_emul.c run->exit_reason = KVM_EXIT_INTERNAL_ERROR; run 477 arch/mips/kvm/trap_emul.c struct kvm_run *run = vcpu->run; run 489 arch/mips/kvm/trap_emul.c er = kvm_mips_emulate_ri_exc(cause, opc, run, vcpu); run 492 arch/mips/kvm/trap_emul.c er = kvm_mips_emulate_msadis_exc(cause, opc, run, vcpu); run 505 arch/mips/kvm/trap_emul.c run->exit_reason = KVM_EXIT_INTERNAL_ERROR; run 1182 arch/mips/kvm/trap_emul.c static void kvm_trap_emul_vcpu_reenter(struct kvm_run *run, run 1226 arch/mips/kvm/trap_emul.c static int kvm_trap_emul_vcpu_run(struct kvm_run *run, struct kvm_vcpu *vcpu) run 1235 arch/mips/kvm/trap_emul.c kvm_trap_emul_vcpu_reenter(run, vcpu); run 1253 arch/mips/kvm/trap_emul.c r = vcpu->arch.vcpu_run(run, vcpu); run 861 arch/mips/kvm/vz.c vcpu->run->exit_reason = KVM_EXIT_INTERNAL_ERROR; run 902 arch/mips/kvm/vz.c struct kvm_run *run, run 1065 arch/mips/kvm/vz.c struct kvm_run *run, run 1137 arch/mips/kvm/vz.c struct kvm_run *run = vcpu->run; run 1153 arch/mips/kvm/vz.c er = kvm_vz_gpsi_cop0(inst, opc, cause, run, vcpu); run 1158 arch/mips/kvm/vz.c er = kvm_vz_gpsi_cache(inst, opc, cause, run, vcpu); run 1166 arch/mips/kvm/vz.c er = kvm_vz_gpsi_cache(inst, opc, cause, run, vcpu); run 1453 arch/mips/kvm/vz.c vcpu->run->exit_reason = KVM_EXIT_INTERNAL_ERROR; run 1468 arch/mips/kvm/vz.c struct kvm_run *run = vcpu->run; run 1496 arch/mips/kvm/vz.c run->exit_reason = KVM_EXIT_INTERNAL_ERROR; run 1515 arch/mips/kvm/vz.c struct kvm_run *run = vcpu->run; run 1527 arch/mips/kvm/vz.c run->exit_reason = KVM_EXIT_INTERNAL_ERROR; run 1538 arch/mips/kvm/vz.c struct kvm_run *run = vcpu->run; run 1549 arch/mips/kvm/vz.c run->exit_reason = KVM_EXIT_INTERNAL_ERROR; run 1558 arch/mips/kvm/vz.c run->exit_reason = KVM_EXIT_INTERNAL_ERROR; run 1563 arch/mips/kvm/vz.c er = kvm_mips_emulate_load(inst, cause, run, vcpu); run 1567 arch/mips/kvm/vz.c run->exit_reason = KVM_EXIT_INTERNAL_ERROR; run 1574 arch/mips/kvm/vz.c run->exit_reason = KVM_EXIT_MMIO; run 1577 arch/mips/kvm/vz.c run->exit_reason = KVM_EXIT_INTERNAL_ERROR; run 1585 arch/mips/kvm/vz.c struct kvm_run *run = vcpu->run; run 1605 arch/mips/kvm/vz.c run->exit_reason = KVM_EXIT_INTERNAL_ERROR; run 1610 arch/mips/kvm/vz.c er = kvm_mips_emulate_store(inst, cause, run, vcpu); run 1614 arch/mips/kvm/vz.c run->exit_reason = KVM_EXIT_INTERNAL_ERROR; run 1621 arch/mips/kvm/vz.c run->exit_reason = KVM_EXIT_MMIO; run 1624 arch/mips/kvm/vz.c run->exit_reason = KVM_EXIT_INTERNAL_ERROR; run 3132 arch/mips/kvm/vz.c static void kvm_vz_vcpu_reenter(struct kvm_run *run, struct kvm_vcpu *vcpu) run 3148 arch/mips/kvm/vz.c static int kvm_vz_vcpu_run(struct kvm_run *run, struct kvm_vcpu *vcpu) run 3161 arch/mips/kvm/vz.c r = vcpu->arch.vcpu_run(run, vcpu); run 158 arch/powerpc/include/asm/kvm_book3s.h extern int kvmppc_book3s_hv_page_fault(struct kvm_run *run, run 163 arch/powerpc/include/asm/kvm_book3s.h extern int kvmppc_hv_emulate_mmio(struct kvm_run *run, struct kvm_vcpu *vcpu, run 177 arch/powerpc/include/asm/kvm_book3s.h extern int kvmppc_book3s_radix_page_fault(struct kvm_run *run, run 237 arch/powerpc/include/asm/kvm_book3s.h extern int kvmppc_emulate_paired_single(struct kvm_run *run, struct kvm_vcpu *vcpu); run 308 arch/powerpc/include/asm/kvm_book3s.h long int kvmhv_nested_page_fault(struct kvm_run *run, struct kvm_vcpu *vcpu); run 66 arch/powerpc/include/asm/kvm_ppc.h extern int kvmppc_handle_load(struct kvm_run *run, struct kvm_vcpu *vcpu, run 69 arch/powerpc/include/asm/kvm_ppc.h extern int kvmppc_handle_loads(struct kvm_run *run, struct kvm_vcpu *vcpu, run 72 arch/powerpc/include/asm/kvm_ppc.h extern int kvmppc_handle_vsx_load(struct kvm_run *run, struct kvm_vcpu *vcpu, run 75 arch/powerpc/include/asm/kvm_ppc.h extern int kvmppc_handle_vmx_load(struct kvm_run *run, struct kvm_vcpu *vcpu, run 77 arch/powerpc/include/asm/kvm_ppc.h extern int kvmppc_handle_vmx_store(struct kvm_run *run, struct kvm_vcpu *vcpu, run 79 arch/powerpc/include/asm/kvm_ppc.h extern int kvmppc_handle_store(struct kvm_run *run, struct kvm_vcpu *vcpu, run 82 arch/powerpc/include/asm/kvm_ppc.h extern int kvmppc_handle_vsx_store(struct kvm_run *run, struct kvm_vcpu *vcpu, run 93 arch/powerpc/include/asm/kvm_ppc.h extern int kvmppc_emulate_instruction(struct kvm_run *run, run 96 arch/powerpc/include/asm/kvm_ppc.h extern int kvmppc_emulate_mmio(struct kvm_run *run, struct kvm_vcpu *vcpu); run 275 arch/powerpc/include/asm/kvm_ppc.h int (*vcpu_run)(struct kvm_run *run, struct kvm_vcpu *vcpu); run 302 arch/powerpc/include/asm/kvm_ppc.h int (*emulate_op)(struct kvm_run *run, struct kvm_vcpu *vcpu, run 20 arch/powerpc/kvm/book3s.h extern int kvmppc_core_emulate_op_pr(struct kvm_run *run, struct kvm_vcpu *vcpu, run 428 arch/powerpc/kvm/book3s_64_mmu_hv.c int kvmppc_hv_emulate_mmio(struct kvm_run *run, struct kvm_vcpu *vcpu, run 488 arch/powerpc/kvm/book3s_64_mmu_hv.c return kvmppc_emulate_mmio(run, vcpu); run 491 arch/powerpc/kvm/book3s_64_mmu_hv.c int kvmppc_book3s_hv_page_fault(struct kvm_run *run, struct kvm_vcpu *vcpu, run 513 arch/powerpc/kvm/book3s_64_mmu_hv.c return kvmppc_book3s_radix_page_fault(run, vcpu, ea, dsisr); run 533 arch/powerpc/kvm/book3s_64_mmu_hv.c return kvmppc_hv_emulate_mmio(run, vcpu, gpa, ea, run 569 arch/powerpc/kvm/book3s_64_mmu_hv.c return kvmppc_hv_emulate_mmio(run, vcpu, gpa, ea, run 889 arch/powerpc/kvm/book3s_64_mmu_radix.c int kvmppc_book3s_radix_page_fault(struct kvm_run *run, struct kvm_vcpu *vcpu, run 932 arch/powerpc/kvm/book3s_64_mmu_radix.c return kvmppc_hv_emulate_mmio(run, vcpu, gpa, ea, writing); run 238 arch/powerpc/kvm/book3s_emulate.c int kvmppc_core_emulate_op_pr(struct kvm_run *run, struct kvm_vcpu *vcpu, run 374 arch/powerpc/kvm/book3s_emulate.c run->papr_hcall.nr = cmd; run 377 arch/powerpc/kvm/book3s_emulate.c run->papr_hcall.args[i] = gpr; run 380 arch/powerpc/kvm/book3s_emulate.c run->exit_reason = KVM_EXIT_PAPR_HCALL; run 632 arch/powerpc/kvm/book3s_emulate.c emulated = kvmppc_emulate_paired_single(run, vcpu); run 1133 arch/powerpc/kvm/book3s_hv.c static int kvmppc_emulate_debug_inst(struct kvm_run *run, run 1148 arch/powerpc/kvm/book3s_hv.c run->exit_reason = KVM_EXIT_DEBUG; run 1149 arch/powerpc/kvm/book3s_hv.c run->debug.arch.address = kvmppc_get_pc(vcpu); run 1250 arch/powerpc/kvm/book3s_hv.c static int kvmppc_handle_exit_hv(struct kvm_run *run, struct kvm_vcpu *vcpu, run 1271 arch/powerpc/kvm/book3s_hv.c run->exit_reason = KVM_EXIT_INTERNAL_ERROR; run 1272 arch/powerpc/kvm/book3s_hv.c run->hw.hardware_exit_reason = vcpu->arch.trap; run 1275 arch/powerpc/kvm/book3s_hv.c run->exit_reason = KVM_EXIT_UNKNOWN; run 1276 arch/powerpc/kvm/book3s_hv.c run->ready_for_interrupt_injection = 1; run 1313 arch/powerpc/kvm/book3s_hv.c run->exit_reason = KVM_EXIT_NMI; run 1314 arch/powerpc/kvm/book3s_hv.c run->hw.hardware_exit_reason = vcpu->arch.trap; run 1316 arch/powerpc/kvm/book3s_hv.c run->flags &= ~KVM_RUN_PPC_NMI_DISP_MASK; run 1319 arch/powerpc/kvm/book3s_hv.c run->flags |= KVM_RUN_PPC_NMI_DISP_FULLY_RECOV; run 1321 arch/powerpc/kvm/book3s_hv.c run->flags |= KVM_RUN_PPC_NMI_DISP_NOT_RECOV; run 1348 arch/powerpc/kvm/book3s_hv.c run->papr_hcall.nr = kvmppc_get_gpr(vcpu, 3); run 1350 arch/powerpc/kvm/book3s_hv.c run->papr_hcall.args[i] = kvmppc_get_gpr(vcpu, 4 + i); run 1351 arch/powerpc/kvm/book3s_hv.c run->exit_reason = KVM_EXIT_PAPR_HCALL; run 1387 arch/powerpc/kvm/book3s_hv.c r = kvmppc_emulate_debug_inst(run, vcpu); run 1431 arch/powerpc/kvm/book3s_hv.c run->hw.hardware_exit_reason = vcpu->arch.trap; run 1439 arch/powerpc/kvm/book3s_hv.c static int kvmppc_handle_nested_exit(struct kvm_run *run, struct kvm_vcpu *vcpu) run 1497 arch/powerpc/kvm/book3s_hv.c r = kvmhv_nested_page_fault(run, vcpu); run 1507 arch/powerpc/kvm/book3s_hv.c r = kvmhv_nested_page_fault(run, vcpu); run 4215 arch/powerpc/kvm/book3s_hv.c static int kvmppc_vcpu_run_hv(struct kvm_run *run, struct kvm_vcpu *vcpu) run 4225 arch/powerpc/kvm/book3s_hv.c run->exit_reason = KVM_EXIT_INTERNAL_ERROR; run 4239 arch/powerpc/kvm/book3s_hv.c run->exit_reason = KVM_EXIT_FAIL_ENTRY; run 4240 arch/powerpc/kvm/book3s_hv.c run->fail_entry.hardware_entry_failure_reason = 0; run 4265 arch/powerpc/kvm/book3s_hv.c run->exit_reason = KVM_EXIT_INTR; run 4300 arch/powerpc/kvm/book3s_hv.c r = kvmhv_run_single_vcpu(run, vcpu, ~(u64)0, run 4303 arch/powerpc/kvm/book3s_hv.c r = kvmppc_run_vcpu(run, vcpu); run 4305 arch/powerpc/kvm/book3s_hv.c if (run->exit_reason == KVM_EXIT_PAPR_HCALL && run 4313 arch/powerpc/kvm/book3s_hv.c r = kvmppc_book3s_hv_page_fault(run, vcpu, run 4967 arch/powerpc/kvm/book3s_hv.c static int kvmppc_core_emulate_op_hv(struct kvm_run *run, struct kvm_vcpu *vcpu, run 1260 arch/powerpc/kvm/book3s_hv_nested.c static long int __kvmhv_nested_page_fault(struct kvm_run *run, run 1344 arch/powerpc/kvm/book3s_hv_nested.c return kvmppc_hv_emulate_mmio(run, vcpu, gpa, ea, writing); run 1431 arch/powerpc/kvm/book3s_hv_nested.c long int kvmhv_nested_page_fault(struct kvm_run *run, struct kvm_vcpu *vcpu) run 1437 arch/powerpc/kvm/book3s_hv_nested.c ret = __kvmhv_nested_page_fault(run, vcpu, gp); run 172 arch/powerpc/kvm/book3s_paired_singles.c static int kvmppc_emulate_fpr_load(struct kvm_run *run, struct kvm_vcpu *vcpu, run 191 arch/powerpc/kvm/book3s_paired_singles.c emulated = kvmppc_handle_load(run, vcpu, KVM_MMIO_REG_FPR | rs, run 216 arch/powerpc/kvm/book3s_paired_singles.c static int kvmppc_emulate_fpr_store(struct kvm_run *run, struct kvm_vcpu *vcpu, run 251 arch/powerpc/kvm/book3s_paired_singles.c emulated = kvmppc_handle_store(run, vcpu, val, len, 1); run 262 arch/powerpc/kvm/book3s_paired_singles.c static int kvmppc_emulate_psq_load(struct kvm_run *run, struct kvm_vcpu *vcpu, run 282 arch/powerpc/kvm/book3s_paired_singles.c emulated = kvmppc_handle_load(run, vcpu, KVM_MMIO_REG_FPR | rs, run 287 arch/powerpc/kvm/book3s_paired_singles.c emulated = kvmppc_handle_load(run, vcpu, KVM_MMIO_REG_FQPR | rs, run 305 arch/powerpc/kvm/book3s_paired_singles.c static int kvmppc_emulate_psq_store(struct kvm_run *run, struct kvm_vcpu *vcpu, run 321 arch/powerpc/kvm/book3s_paired_singles.c emulated = kvmppc_handle_store(run, vcpu, tmp[0], 4, 1); run 324 arch/powerpc/kvm/book3s_paired_singles.c emulated = kvmppc_handle_store(run, vcpu, val, 8, 1); run 621 arch/powerpc/kvm/book3s_paired_singles.c int kvmppc_emulate_paired_single(struct kvm_run *run, struct kvm_vcpu *vcpu) run 683 arch/powerpc/kvm/book3s_paired_singles.c emulated = kvmppc_emulate_psq_load(run, vcpu, ax_rd, addr, w, i); run 693 arch/powerpc/kvm/book3s_paired_singles.c emulated = kvmppc_emulate_psq_load(run, vcpu, ax_rd, addr, w, i); run 706 arch/powerpc/kvm/book3s_paired_singles.c emulated = kvmppc_emulate_psq_store(run, vcpu, ax_rd, addr, w, i); run 716 arch/powerpc/kvm/book3s_paired_singles.c emulated = kvmppc_emulate_psq_store(run, vcpu, ax_rd, addr, w, i); run 736 arch/powerpc/kvm/book3s_paired_singles.c emulated = kvmppc_emulate_psq_load(run, vcpu, ax_rd, addr, w, i); run 750 arch/powerpc/kvm/book3s_paired_singles.c emulated = kvmppc_emulate_psq_load(run, vcpu, ax_rd, addr, w, i); run 827 arch/powerpc/kvm/book3s_paired_singles.c emulated = kvmppc_emulate_psq_store(run, vcpu, ax_rd, addr, w, i); run 837 arch/powerpc/kvm/book3s_paired_singles.c emulated = kvmppc_emulate_psq_store(run, vcpu, ax_rd, addr, w, i); run 925 arch/powerpc/kvm/book3s_paired_singles.c emulated = kvmppc_emulate_fpr_load(run, vcpu, ax_rd, addr, run 933 arch/powerpc/kvm/book3s_paired_singles.c emulated = kvmppc_emulate_fpr_load(run, vcpu, ax_rd, addr, run 944 arch/powerpc/kvm/book3s_paired_singles.c emulated = kvmppc_emulate_fpr_load(run, vcpu, ax_rd, addr, run 952 arch/powerpc/kvm/book3s_paired_singles.c emulated = kvmppc_emulate_fpr_load(run, vcpu, ax_rd, addr, run 963 arch/powerpc/kvm/book3s_paired_singles.c emulated = kvmppc_emulate_fpr_store(run, vcpu, ax_rd, addr, run 971 arch/powerpc/kvm/book3s_paired_singles.c emulated = kvmppc_emulate_fpr_store(run, vcpu, ax_rd, addr, run 982 arch/powerpc/kvm/book3s_paired_singles.c emulated = kvmppc_emulate_fpr_store(run, vcpu, ax_rd, addr, run 990 arch/powerpc/kvm/book3s_paired_singles.c emulated = kvmppc_emulate_fpr_store(run, vcpu, ax_rd, addr, run 1004 arch/powerpc/kvm/book3s_paired_singles.c emulated = kvmppc_emulate_fpr_load(run, vcpu, ax_rd, run 1013 arch/powerpc/kvm/book3s_paired_singles.c emulated = kvmppc_emulate_fpr_load(run, vcpu, ax_rd, run 1025 arch/powerpc/kvm/book3s_paired_singles.c emulated = kvmppc_emulate_fpr_load(run, vcpu, ax_rd, run 1034 arch/powerpc/kvm/book3s_paired_singles.c emulated = kvmppc_emulate_fpr_load(run, vcpu, ax_rd, run 1046 arch/powerpc/kvm/book3s_paired_singles.c emulated = kvmppc_emulate_fpr_store(run, vcpu, ax_rd, run 1055 arch/powerpc/kvm/book3s_paired_singles.c emulated = kvmppc_emulate_fpr_store(run, vcpu, ax_rd, run 1067 arch/powerpc/kvm/book3s_paired_singles.c emulated = kvmppc_emulate_fpr_store(run, vcpu, ax_rd, run 1076 arch/powerpc/kvm/book3s_paired_singles.c emulated = kvmppc_emulate_fpr_store(run, vcpu, ax_rd, run 1088 arch/powerpc/kvm/book3s_paired_singles.c emulated = kvmppc_emulate_fpr_store(run, vcpu, ax_rd, run 667 arch/powerpc/kvm/book3s_pr.c int kvmppc_handle_pagefault(struct kvm_run *run, struct kvm_vcpu *vcpu, run 762 arch/powerpc/kvm/book3s_pr.c run->exit_reason = KVM_EXIT_INTERNAL_ERROR; run 775 arch/powerpc/kvm/book3s_pr.c r = kvmppc_emulate_mmio(run, vcpu); run 959 arch/powerpc/kvm/book3s_pr.c er = kvmppc_emulate_instruction(vcpu->run, vcpu); run 1056 arch/powerpc/kvm/book3s_pr.c static int kvmppc_exit_pr_progint(struct kvm_run *run, struct kvm_vcpu *vcpu, run 1091 arch/powerpc/kvm/book3s_pr.c er = kvmppc_emulate_instruction(run, vcpu); run 1106 arch/powerpc/kvm/book3s_pr.c run->exit_reason = KVM_EXIT_MMIO; run 1119 arch/powerpc/kvm/book3s_pr.c int kvmppc_handle_exit_pr(struct kvm_run *run, struct kvm_vcpu *vcpu, run 1127 arch/powerpc/kvm/book3s_pr.c run->exit_reason = KVM_EXIT_UNKNOWN; run 1128 arch/powerpc/kvm/book3s_pr.c run->ready_for_interrupt_injection = 1; run 1165 arch/powerpc/kvm/book3s_pr.c r = kvmppc_handle_pagefault(run, vcpu, kvmppc_get_pc(vcpu), exit_nr); run 1215 arch/powerpc/kvm/book3s_pr.c r = kvmppc_handle_pagefault(run, vcpu, dar, exit_nr); run 1259 arch/powerpc/kvm/book3s_pr.c r = kvmppc_exit_pr_progint(run, vcpu, exit_nr); run 1291 arch/powerpc/kvm/book3s_pr.c run->papr_hcall.nr = cmd; run 1294 arch/powerpc/kvm/book3s_pr.c run->papr_hcall.args[i] = gpr; run 1296 arch/powerpc/kvm/book3s_pr.c run->exit_reason = KVM_EXIT_PAPR_HCALL; run 1303 arch/powerpc/kvm/book3s_pr.c u64 *gprs = run->osi.gprs; run 1306 arch/powerpc/kvm/book3s_pr.c run->exit_reason = KVM_EXIT_OSI; run 1337 arch/powerpc/kvm/book3s_pr.c r = kvmppc_exit_pr_progint(run, vcpu, exit_nr); run 1393 arch/powerpc/kvm/book3s_pr.c run->exit_reason = KVM_EXIT_DEBUG; run 719 arch/powerpc/kvm/booke.c vcpu->run->exit_reason = KVM_EXIT_WATCHDOG; run 724 arch/powerpc/kvm/booke.c vcpu->run->epr.epr = 0; run 726 arch/powerpc/kvm/booke.c vcpu->run->exit_reason = KVM_EXIT_EPR; run 803 arch/powerpc/kvm/booke.c static int emulation_exit(struct kvm_run *run, struct kvm_vcpu *vcpu) run 807 arch/powerpc/kvm/booke.c er = kvmppc_emulate_instruction(run, vcpu); run 824 arch/powerpc/kvm/booke.c run->hw.hardware_exit_reason = ~0ULL << 32; run 825 arch/powerpc/kvm/booke.c run->hw.hardware_exit_reason |= vcpu->arch.last_inst; run 837 arch/powerpc/kvm/booke.c static int kvmppc_handle_debug(struct kvm_run *run, struct kvm_vcpu *vcpu) run 869 arch/powerpc/kvm/booke.c run->debug.arch.status = 0; run 870 arch/powerpc/kvm/booke.c run->debug.arch.address = vcpu->arch.regs.nip; run 873 arch/powerpc/kvm/booke.c run->debug.arch.status |= KVMPPC_DEBUG_BREAKPOINT; run 876 arch/powerpc/kvm/booke.c run->debug.arch.status |= KVMPPC_DEBUG_WATCH_WRITE; run 878 arch/powerpc/kvm/booke.c run->debug.arch.status |= KVMPPC_DEBUG_WATCH_READ; run 880 arch/powerpc/kvm/booke.c run->debug.arch.address = dbg_reg->dac1; run 882 arch/powerpc/kvm/booke.c run->debug.arch.address = dbg_reg->dac2; run 957 arch/powerpc/kvm/booke.c static int kvmppc_resume_inst_load(struct kvm_run *run, struct kvm_vcpu *vcpu, run 969 arch/powerpc/kvm/booke.c run->hw.hardware_exit_reason = ~0ULL << 32; run 970 arch/powerpc/kvm/booke.c run->hw.hardware_exit_reason |= last_inst; run 984 arch/powerpc/kvm/booke.c int kvmppc_handle_exit(struct kvm_run *run, struct kvm_vcpu *vcpu, run 1023 arch/powerpc/kvm/booke.c run->exit_reason = KVM_EXIT_UNKNOWN; run 1024 arch/powerpc/kvm/booke.c run->ready_for_interrupt_injection = 1; run 1027 arch/powerpc/kvm/booke.c r = kvmppc_resume_inst_load(run, vcpu, emulated, last_inst); run 1036 arch/powerpc/kvm/booke.c run->hw.hardware_exit_reason = ~1ULL << 32; run 1037 arch/powerpc/kvm/booke.c run->hw.hardware_exit_reason |= mfspr(SPRN_MCSR); run 1087 arch/powerpc/kvm/booke.c r = emulation_exit(run, vcpu); run 1097 arch/powerpc/kvm/booke.c r = kvmppc_handle_debug(run, vcpu); run 1098 arch/powerpc/kvm/booke.c run->exit_reason = KVM_EXIT_DEBUG; run 1118 arch/powerpc/kvm/booke.c r = emulation_exit(run, vcpu); run 1165 arch/powerpc/kvm/booke.c run->hw.hardware_exit_reason = exit_nr; run 1285 arch/powerpc/kvm/booke.c r = kvmppc_emulate_mmio(run, vcpu); run 1336 arch/powerpc/kvm/booke.c r = kvmppc_handle_debug(run, vcpu); run 1338 arch/powerpc/kvm/booke.c run->exit_reason = KVM_EXIT_DEBUG; run 73 arch/powerpc/kvm/booke.h int kvmppc_booke_emulate_op(struct kvm_run *run, struct kvm_vcpu *vcpu, run 98 arch/powerpc/kvm/booke.h extern int kvmppc_core_emulate_op_e500(struct kvm_run *run, run 106 arch/powerpc/kvm/booke.h extern int kvmppc_core_emulate_op_e500(struct kvm_run *run, run 42 arch/powerpc/kvm/booke_emulate.c int kvmppc_booke_emulate_op(struct kvm_run *run, struct kvm_vcpu *vcpu, run 86 arch/powerpc/kvm/e500_emulate.c static int kvmppc_e500_emul_ehpriv(struct kvm_run *run, struct kvm_vcpu *vcpu, run 93 arch/powerpc/kvm/e500_emulate.c run->exit_reason = KVM_EXIT_DEBUG; run 94 arch/powerpc/kvm/e500_emulate.c run->debug.arch.address = vcpu->arch.regs.nip; run 95 arch/powerpc/kvm/e500_emulate.c run->debug.arch.status = 0; run 128 arch/powerpc/kvm/e500_emulate.c int kvmppc_core_emulate_op_e500(struct kvm_run *run, struct kvm_vcpu *vcpu, run 185 arch/powerpc/kvm/e500_emulate.c emulated = kvmppc_e500_emul_ehpriv(run, vcpu, inst, run 200 arch/powerpc/kvm/e500_emulate.c emulated = kvmppc_booke_emulate_op(run, vcpu, inst, advance); run 194 arch/powerpc/kvm/emulate.c int kvmppc_emulate_instruction(struct kvm_run *run, struct kvm_vcpu *vcpu) run 273 arch/powerpc/kvm/emulate.c run->exit_reason = KVM_EXIT_DEBUG; run 274 arch/powerpc/kvm/emulate.c run->debug.arch.status = 0; run 275 arch/powerpc/kvm/emulate.c run->debug.arch.address = kvmppc_get_pc(vcpu); run 288 arch/powerpc/kvm/emulate.c emulated = vcpu->kvm->arch.kvm_ops->emulate_op(run, vcpu, inst, run 74 arch/powerpc/kvm/emulate_loadstore.c struct kvm_run *run = vcpu->run; run 107 arch/powerpc/kvm/emulate_loadstore.c emulated = kvmppc_handle_loads(run, vcpu, run 110 arch/powerpc/kvm/emulate_loadstore.c emulated = kvmppc_handle_load(run, vcpu, run 127 arch/powerpc/kvm/emulate_loadstore.c emulated = kvmppc_handle_loads(run, vcpu, run 130 arch/powerpc/kvm/emulate_loadstore.c emulated = kvmppc_handle_load(run, vcpu, run 167 arch/powerpc/kvm/emulate_loadstore.c emulated = kvmppc_handle_vmx_load(run, run 172 arch/powerpc/kvm/emulate_loadstore.c emulated = kvmppc_handle_vmx_load(run, vcpu, run 220 arch/powerpc/kvm/emulate_loadstore.c emulated = kvmppc_handle_vsx_load(run, vcpu, run 230 arch/powerpc/kvm/emulate_loadstore.c emulated = kvmppc_handle_store(run, vcpu, op.val, run 253 arch/powerpc/kvm/emulate_loadstore.c emulated = kvmppc_handle_store(run, vcpu, run 293 arch/powerpc/kvm/emulate_loadstore.c emulated = kvmppc_handle_vmx_store(run, run 297 arch/powerpc/kvm/emulate_loadstore.c emulated = kvmppc_handle_vmx_store(run, run 341 arch/powerpc/kvm/emulate_loadstore.c emulated = kvmppc_handle_vsx_store(run, vcpu, run 94 arch/powerpc/kvm/powerpc.c vcpu->run->exit_reason = KVM_EXIT_INTR; run 281 arch/powerpc/kvm/powerpc.c int kvmppc_emulate_mmio(struct kvm_run *run, struct kvm_vcpu *vcpu) run 297 arch/powerpc/kvm/powerpc.c run->exit_reason = KVM_EXIT_MMIO; run 1111 arch/powerpc/kvm/powerpc.c struct kvm_run *run) run 1115 arch/powerpc/kvm/powerpc.c if (run->mmio.len > sizeof(gpr)) { run 1116 arch/powerpc/kvm/powerpc.c printk(KERN_ERR "bad MMIO length: %d\n", run->mmio.len); run 1121 arch/powerpc/kvm/powerpc.c switch (run->mmio.len) { run 1122 arch/powerpc/kvm/powerpc.c case 8: gpr = *(u64 *)run->mmio.data; break; run 1123 arch/powerpc/kvm/powerpc.c case 4: gpr = *(u32 *)run->mmio.data; break; run 1124 arch/powerpc/kvm/powerpc.c case 2: gpr = *(u16 *)run->mmio.data; break; run 1125 arch/powerpc/kvm/powerpc.c case 1: gpr = *(u8 *)run->mmio.data; break; run 1128 arch/powerpc/kvm/powerpc.c switch (run->mmio.len) { run 1129 arch/powerpc/kvm/powerpc.c case 8: gpr = swab64(*(u64 *)run->mmio.data); break; run 1130 arch/powerpc/kvm/powerpc.c case 4: gpr = swab32(*(u32 *)run->mmio.data); break; run 1131 arch/powerpc/kvm/powerpc.c case 2: gpr = swab16(*(u16 *)run->mmio.data); break; run 1132 arch/powerpc/kvm/powerpc.c case 1: gpr = *(u8 *)run->mmio.data; break; run 1137 arch/powerpc/kvm/powerpc.c if ((vcpu->arch.mmio_sp64_extend) && (run->mmio.len == 4)) run 1141 arch/powerpc/kvm/powerpc.c switch (run->mmio.len) { run 1222 arch/powerpc/kvm/powerpc.c static int __kvmppc_handle_load(struct kvm_run *run, struct kvm_vcpu *vcpu, run 1236 arch/powerpc/kvm/powerpc.c if (bytes > sizeof(run->mmio.data)) { run 1238 arch/powerpc/kvm/powerpc.c run->mmio.len); run 1241 arch/powerpc/kvm/powerpc.c run->mmio.phys_addr = vcpu->arch.paddr_accessed; run 1242 arch/powerpc/kvm/powerpc.c run->mmio.len = bytes; run 1243 arch/powerpc/kvm/powerpc.c run->mmio.is_write = 0; run 1253 arch/powerpc/kvm/powerpc.c ret = kvm_io_bus_read(vcpu, KVM_MMIO_BUS, run->mmio.phys_addr, run 1254 arch/powerpc/kvm/powerpc.c bytes, &run->mmio.data); run 1259 arch/powerpc/kvm/powerpc.c kvmppc_complete_mmio_load(vcpu, run); run 1267 arch/powerpc/kvm/powerpc.c int kvmppc_handle_load(struct kvm_run *run, struct kvm_vcpu *vcpu, run 1271 arch/powerpc/kvm/powerpc.c return __kvmppc_handle_load(run, vcpu, rt, bytes, is_default_endian, 0); run 1276 arch/powerpc/kvm/powerpc.c int kvmppc_handle_loads(struct kvm_run *run, struct kvm_vcpu *vcpu, run 1280 arch/powerpc/kvm/powerpc.c return __kvmppc_handle_load(run, vcpu, rt, bytes, is_default_endian, 1); run 1284 arch/powerpc/kvm/powerpc.c int kvmppc_handle_vsx_load(struct kvm_run *run, struct kvm_vcpu *vcpu, run 1295 arch/powerpc/kvm/powerpc.c emulated = __kvmppc_handle_load(run, vcpu, rt, bytes, run 1301 arch/powerpc/kvm/powerpc.c vcpu->arch.paddr_accessed += run->mmio.len; run 1310 arch/powerpc/kvm/powerpc.c int kvmppc_handle_store(struct kvm_run *run, struct kvm_vcpu *vcpu, run 1313 arch/powerpc/kvm/powerpc.c void *data = run->mmio.data; run 1324 arch/powerpc/kvm/powerpc.c if (bytes > sizeof(run->mmio.data)) { run 1326 arch/powerpc/kvm/powerpc.c run->mmio.len); run 1329 arch/powerpc/kvm/powerpc.c run->mmio.phys_addr = vcpu->arch.paddr_accessed; run 1330 arch/powerpc/kvm/powerpc.c run->mmio.len = bytes; run 1331 arch/powerpc/kvm/powerpc.c run->mmio.is_write = 1; run 1357 arch/powerpc/kvm/powerpc.c ret = kvm_io_bus_write(vcpu, KVM_MMIO_BUS, run->mmio.phys_addr, run 1358 arch/powerpc/kvm/powerpc.c bytes, &run->mmio.data); run 1426 arch/powerpc/kvm/powerpc.c int kvmppc_handle_vsx_store(struct kvm_run *run, struct kvm_vcpu *vcpu, run 1442 arch/powerpc/kvm/powerpc.c emulated = kvmppc_handle_store(run, vcpu, run 1448 arch/powerpc/kvm/powerpc.c vcpu->arch.paddr_accessed += run->mmio.len; run 1458 arch/powerpc/kvm/powerpc.c struct kvm_run *run) run 1463 arch/powerpc/kvm/powerpc.c vcpu->arch.paddr_accessed += run->mmio.len; run 1466 arch/powerpc/kvm/powerpc.c emulated = kvmppc_handle_vsx_load(run, vcpu, vcpu->arch.io_gpr, run 1467 arch/powerpc/kvm/powerpc.c run->mmio.len, 1, vcpu->arch.mmio_sign_extend); run 1469 arch/powerpc/kvm/powerpc.c emulated = kvmppc_handle_vsx_store(run, vcpu, run 1470 arch/powerpc/kvm/powerpc.c vcpu->arch.io_gpr, run->mmio.len, 1); run 1475 arch/powerpc/kvm/powerpc.c run->exit_reason = KVM_EXIT_MMIO; run 1480 arch/powerpc/kvm/powerpc.c run->exit_reason = KVM_EXIT_INTERNAL_ERROR; run 1481 arch/powerpc/kvm/powerpc.c run->internal.suberror = KVM_INTERNAL_ERROR_EMULATION; run 1493 arch/powerpc/kvm/powerpc.c int kvmppc_handle_vmx_load(struct kvm_run *run, struct kvm_vcpu *vcpu, run 1502 arch/powerpc/kvm/powerpc.c emulated = __kvmppc_handle_load(run, vcpu, rt, bytes, run 1508 arch/powerpc/kvm/powerpc.c vcpu->arch.paddr_accessed += run->mmio.len; run 1588 arch/powerpc/kvm/powerpc.c int kvmppc_handle_vmx_store(struct kvm_run *run, struct kvm_vcpu *vcpu, run 1623 arch/powerpc/kvm/powerpc.c emulated = kvmppc_handle_store(run, vcpu, val, bytes, run 1628 arch/powerpc/kvm/powerpc.c vcpu->arch.paddr_accessed += run->mmio.len; run 1637 arch/powerpc/kvm/powerpc.c struct kvm_run *run) run 1642 arch/powerpc/kvm/powerpc.c vcpu->arch.paddr_accessed += run->mmio.len; run 1645 arch/powerpc/kvm/powerpc.c emulated = kvmppc_handle_vmx_load(run, vcpu, run 1646 arch/powerpc/kvm/powerpc.c vcpu->arch.io_gpr, run->mmio.len, 1); run 1648 arch/powerpc/kvm/powerpc.c emulated = kvmppc_handle_vmx_store(run, vcpu, run 1649 arch/powerpc/kvm/powerpc.c vcpu->arch.io_gpr, run->mmio.len, 1); run 1654 arch/powerpc/kvm/powerpc.c run->exit_reason = KVM_EXIT_MMIO; run 1659 arch/powerpc/kvm/powerpc.c run->exit_reason = KVM_EXIT_INTERNAL_ERROR; run 1660 arch/powerpc/kvm/powerpc.c run->internal.suberror = KVM_INTERNAL_ERROR_EMULATION; run 1768 arch/powerpc/kvm/powerpc.c int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu, struct kvm_run *run) run 1777 arch/powerpc/kvm/powerpc.c kvmppc_complete_mmio_load(vcpu, run); run 1785 arch/powerpc/kvm/powerpc.c r = kvmppc_emulate_mmio_vsx_loadstore(vcpu, run); run 1799 arch/powerpc/kvm/powerpc.c r = kvmppc_emulate_mmio_vmx_loadstore(vcpu, run); run 1807 arch/powerpc/kvm/powerpc.c u64 *gprs = run->osi.gprs; run 1816 arch/powerpc/kvm/powerpc.c kvmppc_set_gpr(vcpu, 3, run->papr_hcall.ret); run 1818 arch/powerpc/kvm/powerpc.c kvmppc_set_gpr(vcpu, 4 + i, run->papr_hcall.args[i]); run 1822 arch/powerpc/kvm/powerpc.c kvmppc_set_epr(vcpu, run->epr.epr); run 1829 arch/powerpc/kvm/powerpc.c if (run->immediate_exit) run 1832 arch/powerpc/kvm/powerpc.c r = kvmppc_vcpu_run(run, vcpu); run 475 arch/powerpc/kvm/trace_hv.h TP_PROTO(struct kvm_vcpu *vcpu, struct kvm_run *run), run 477 arch/powerpc/kvm/trace_hv.h TP_ARGS(vcpu, run), run 487 arch/powerpc/kvm/trace_hv.h __entry->exit = run->exit_reason; run 26 arch/s390/kvm/diag.c start = vcpu->run->s.regs.gprs[(vcpu->arch.sie_block->ipa & 0xf0) >> 4]; run 27 arch/s390/kvm/diag.c end = vcpu->run->s.regs.gprs[vcpu->arch.sie_block->ipa & 0xf] + PAGE_SIZE; run 77 arch/s390/kvm/diag.c vcpu->run->s.regs.gprs[rx]); run 79 arch/s390/kvm/diag.c if (vcpu->run->s.regs.gprs[rx] & 7) run 81 arch/s390/kvm/diag.c rc = read_guest(vcpu, vcpu->run->s.regs.gprs[rx], rx, &parm, sizeof(parm)); run 98 arch/s390/kvm/diag.c vcpu->run->s.regs.gprs[ry] = 8; run 112 arch/s390/kvm/diag.c vcpu->run->s.regs.gprs[ry] = 0; run 126 arch/s390/kvm/diag.c vcpu->run->s.regs.gprs[ry] = 0; run 132 arch/s390/kvm/diag.c vcpu->run->s.regs.gprs[ry] = 4; run 159 arch/s390/kvm/diag.c tid = vcpu->run->s.regs.gprs[(vcpu->arch.sie_block->ipa & 0xf0) >> 4]; run 175 arch/s390/kvm/diag.c unsigned long subcode = vcpu->run->s.regs.gprs[reg] & 0xffff; run 181 arch/s390/kvm/diag.c vcpu->run->s390_reset_flags = KVM_S390_RESET_CLEAR; run 184 arch/s390/kvm/diag.c vcpu->run->s390_reset_flags = 0; run 192 arch/s390/kvm/diag.c vcpu->run->s390_reset_flags |= KVM_S390_RESET_SUBSYSTEM; run 193 arch/s390/kvm/diag.c vcpu->run->s390_reset_flags |= KVM_S390_RESET_IPL; run 194 arch/s390/kvm/diag.c vcpu->run->s390_reset_flags |= KVM_S390_RESET_CPU_INIT; run 195 arch/s390/kvm/diag.c vcpu->run->exit_reason = KVM_EXIT_S390_RESET; run 197 arch/s390/kvm/diag.c vcpu->run->s390_reset_flags); run 198 arch/s390/kvm/diag.c trace_kvm_s390_request_resets(vcpu->run->s390_reset_flags); run 209 arch/s390/kvm/diag.c (vcpu->run->s.regs.gprs[1] != KVM_S390_VIRTIO_CCW_NOTIFY)) run 213 arch/s390/kvm/diag.c (u32) vcpu->run->s.regs.gprs[2], run 214 arch/s390/kvm/diag.c (u32) vcpu->run->s.regs.gprs[3], run 215 arch/s390/kvm/diag.c vcpu->run->s.regs.gprs[4]); run 224 arch/s390/kvm/diag.c vcpu->run->s.regs.gprs[2] & 0xffffffff, run 225 arch/s390/kvm/diag.c 8, &vcpu->run->s.regs.gprs[3], run 226 arch/s390/kvm/diag.c vcpu->run->s.regs.gprs[4]); run 233 arch/s390/kvm/diag.c vcpu->run->s.regs.gprs[2] = ret; run 393 arch/s390/kvm/gaccess.c save_access_regs(vcpu->run->s.regs.acrs); run 394 arch/s390/kvm/gaccess.c alet.val = vcpu->run->s.regs.acrs[ar]; run 373 arch/s390/kvm/guestdbg.c vcpu->run->exit_reason = KVM_EXIT_DEBUG; run 391 arch/s390/kvm/guestdbg.c struct kvm_debug_exit_arch *debug_exit = &vcpu->run->debug.arch; run 485 arch/s390/kvm/guestdbg.c *addr = base ? vcpu->run->s.regs.gprs[base] : 0; run 486 arch/s390/kvm/guestdbg.c *addr += index ? vcpu->run->s.regs.gprs[index] : 0; run 322 arch/s390/kvm/intercept.c rc = guest_translate_address(vcpu, vcpu->run->s.regs.gprs[reg2], run 331 arch/s390/kvm/intercept.c rc = guest_translate_address(vcpu, vcpu->run->s.regs.gprs[reg1], run 371 arch/s390/kvm/intercept.c code = vcpu->run->s.regs.gprs[reg1]; run 372 arch/s390/kvm/intercept.c addr = vcpu->run->s.regs.gprs[reg2]; run 406 arch/s390/kvm/intercept.c vcpu->run->s.regs.gprs[reg2 + 1] = rc; run 559 arch/s390/kvm/interrupt.c save_access_regs(vcpu->run->s.regs.acrs); run 589 arch/s390/kvm/interrupt.c if (write_guest_abs(vcpu, ext_sa_addr, vcpu->run->s.regs.vrs, run 598 arch/s390/kvm/interrupt.c &vcpu->run->s.regs.gscb, 32)) run 614 arch/s390/kvm/interrupt.c convert_vx_to_fp(fprs, (__vector128 *) vcpu->run->s.regs.vrs); run 618 arch/s390/kvm/interrupt.c vcpu->run->s.regs.fprs, 128); run 621 arch/s390/kvm/interrupt.c vcpu->run->s.regs.gprs, 128); run 631 arch/s390/kvm/interrupt.c &vcpu->run->s.regs.acrs, 64); run 2716 arch/s390/kvm/kvm-s390.c vcpu->run->kvm_valid_regs = KVM_SYNC_PREFIX | run 2724 arch/s390/kvm/kvm-s390.c vcpu->run->kvm_valid_regs |= KVM_SYNC_RICCB; run 2726 arch/s390/kvm/kvm-s390.c vcpu->run->kvm_valid_regs |= KVM_SYNC_BPBC; run 2728 arch/s390/kvm/kvm-s390.c vcpu->run->kvm_valid_regs |= KVM_SYNC_GSCB; run 2730 arch/s390/kvm/kvm-s390.c vcpu->run->kvm_valid_regs |= KVM_SYNC_ETOKEN; run 2735 arch/s390/kvm/kvm-s390.c vcpu->run->kvm_valid_regs |= KVM_SYNC_VRS; run 2737 arch/s390/kvm/kvm-s390.c vcpu->run->kvm_valid_regs |= KVM_SYNC_FPRS; run 2869 arch/s390/kvm/kvm-s390.c vcpu->run->s.regs.fpc = 0; run 3018 arch/s390/kvm/kvm-s390.c vcpu->arch.sie_block->sdnxo = ((unsigned long) &vcpu->run->s.regs.sdnx) run 3020 arch/s390/kvm/kvm-s390.c vcpu->arch.sie_block->riccbd = (unsigned long) &vcpu->run->s.regs.riccb; run 3303 arch/s390/kvm/kvm-s390.c memcpy(&vcpu->run->s.regs.gprs, ®s->gprs, sizeof(regs->gprs)); run 3311 arch/s390/kvm/kvm-s390.c memcpy(®s->gprs, &vcpu->run->s.regs.gprs, sizeof(regs->gprs)); run 3321 arch/s390/kvm/kvm-s390.c memcpy(&vcpu->run->s.regs.acrs, &sregs->acrs, sizeof(sregs->acrs)); run 3333 arch/s390/kvm/kvm-s390.c memcpy(&sregs->acrs, &vcpu->run->s.regs.acrs, sizeof(sregs->acrs)); run 3350 arch/s390/kvm/kvm-s390.c vcpu->run->s.regs.fpc = fpu->fpc; run 3352 arch/s390/kvm/kvm-s390.c convert_fp_to_vx((__vector128 *) vcpu->run->s.regs.vrs, run 3355 arch/s390/kvm/kvm-s390.c memcpy(vcpu->run->s.regs.fprs, &fpu->fprs, sizeof(fpu->fprs)); run 3370 arch/s390/kvm/kvm-s390.c (__vector128 *) vcpu->run->s.regs.vrs); run 3372 arch/s390/kvm/kvm-s390.c memcpy(fpu->fprs, vcpu->run->s.regs.fprs, sizeof(fpu->fprs)); run 3373 arch/s390/kvm/kvm-s390.c fpu->fpc = vcpu->run->s.regs.fpc; run 3386 arch/s390/kvm/kvm-s390.c vcpu->run->psw_mask = psw.mask; run 3387 arch/s390/kvm/kvm-s390.c vcpu->run->psw_addr = psw.addr; run 3705 arch/s390/kvm/kvm-s390.c vcpu->arch.sie_block->gg14 = vcpu->run->s.regs.gprs[14]; run 3706 arch/s390/kvm/kvm-s390.c vcpu->arch.sie_block->gg15 = vcpu->run->s.regs.gprs[15]; run 3787 arch/s390/kvm/kvm-s390.c vcpu->run->s.regs.gprs[14] = vcpu->arch.sie_block->gg14; run 3788 arch/s390/kvm/kvm-s390.c vcpu->run->s.regs.gprs[15] = vcpu->arch.sie_block->gg15; run 3804 arch/s390/kvm/kvm-s390.c vcpu->run->exit_reason = KVM_EXIT_S390_SIEIC; run 3805 arch/s390/kvm/kvm-s390.c vcpu->run->s390_sieic.icptcode = vcpu->arch.sie_block->icptcode; run 3806 arch/s390/kvm/kvm-s390.c vcpu->run->s390_sieic.ipa = vcpu->arch.sie_block->ipa; run 3807 arch/s390/kvm/kvm-s390.c vcpu->run->s390_sieic.ipb = vcpu->arch.sie_block->ipb; run 3813 arch/s390/kvm/kvm-s390.c vcpu->run->exit_reason = KVM_EXIT_S390_UCONTROL; run 3814 arch/s390/kvm/kvm-s390.c vcpu->run->s390_ucontrol.trans_exc_code = run 3816 arch/s390/kvm/kvm-s390.c vcpu->run->s390_ucontrol.pgm_code = 0x10; run 3853 arch/s390/kvm/kvm-s390.c vcpu->run->s.regs.gprs); run 3927 arch/s390/kvm/kvm-s390.c restore_access_regs(vcpu->run->s.regs.acrs); run 3933 arch/s390/kvm/kvm-s390.c current->thread.fpu.regs = vcpu->run->s.regs.vrs; run 3935 arch/s390/kvm/kvm-s390.c current->thread.fpu.regs = vcpu->run->s.regs.fprs; run 3936 arch/s390/kvm/kvm-s390.c current->thread.fpu.fpc = vcpu->run->s.regs.fpc; run 3949 arch/s390/kvm/kvm-s390.c &vcpu->run->s.regs.gscb; run 3974 arch/s390/kvm/kvm-s390.c save_access_regs(vcpu->run->s.regs.acrs); run 3978 arch/s390/kvm/kvm-s390.c vcpu->run->s.regs.fpc = current->thread.fpu.fpc; run 4087 arch/s390/kvm/kvm-s390.c convert_vx_to_fp(fprs, (__vector128 *) vcpu->run->s.regs.vrs); run 4092 arch/s390/kvm/kvm-s390.c vcpu->run->s.regs.fprs, 128); run 4095 arch/s390/kvm/kvm-s390.c vcpu->run->s.regs.gprs, 128); run 4101 arch/s390/kvm/kvm-s390.c &vcpu->run->s.regs.fpc, 4); run 4111 arch/s390/kvm/kvm-s390.c &vcpu->run->s.regs.acrs, 64); run 4125 arch/s390/kvm/kvm-s390.c vcpu->run->s.regs.fpc = current->thread.fpu.fpc; run 4126 arch/s390/kvm/kvm-s390.c save_access_regs(vcpu->run->s.regs.acrs); run 107 arch/s390/kvm/kvm-s390.h return (base2 ? vcpu->run->s.regs.gprs[base2] : 0) + disp2; run 119 arch/s390/kvm/kvm-s390.h *address1 = (base1 ? vcpu->run->s.regs.gprs[base1] : 0) + disp1; run 120 arch/s390/kvm/kvm-s390.h *address2 = (base2 ? vcpu->run->s.regs.gprs[base2] : 0) + disp2; run 148 arch/s390/kvm/kvm-s390.h return (base2 ? vcpu->run->s.regs.gprs[base2] : 0) + (long)(int)disp2; run 159 arch/s390/kvm/kvm-s390.h return (base2 ? vcpu->run->s.regs.gprs[base2] : 0) + disp2; run 64 arch/s390/kvm/priv.c current->thread.gs_cb = (struct gs_cb *)&vcpu->run->s.regs.gscb; run 265 arch/s390/kvm/priv.c gaddr = vcpu->run->s.regs.gprs[reg2] & PAGE_MASK; run 289 arch/s390/kvm/priv.c vcpu->run->s.regs.gprs[reg1] &= ~0xff; run 290 arch/s390/kvm/priv.c vcpu->run->s.regs.gprs[reg1] |= key; run 312 arch/s390/kvm/priv.c gaddr = vcpu->run->s.regs.gprs[reg2] & PAGE_MASK; run 370 arch/s390/kvm/priv.c key = vcpu->run->s.regs.gprs[reg1] & 0xfe; run 371 arch/s390/kvm/priv.c start = vcpu->run->s.regs.gprs[reg2] & PAGE_MASK; run 412 arch/s390/kvm/priv.c vcpu->run->s.regs.gprs[reg1] &= ~0xff00UL; run 413 arch/s390/kvm/priv.c vcpu->run->s.regs.gprs[reg1] |= (u64) oldkey << 8; run 418 arch/s390/kvm/priv.c vcpu->run->s.regs.gprs[reg2] &= ~PAGE_MASK; run 420 arch/s390/kvm/priv.c vcpu->run->s.regs.gprs[reg2] &= ~0xfffff000UL; run 422 arch/s390/kvm/priv.c vcpu->run->s.regs.gprs[reg2] |= end; run 449 arch/s390/kvm/priv.c addr = vcpu->run->s.regs.gprs[reg2] & PAGE_MASK; run 464 arch/s390/kvm/priv.c vcpu->run->s.regs.gprs[0] = 0; run 542 arch/s390/kvm/priv.c if (vcpu->run->s.regs.gprs[1]) run 544 arch/s390/kvm/priv.c vcpu->run->s.regs.gprs[1]); run 554 arch/s390/kvm/priv.c vcpu->run->exit_reason = KVM_EXIT_S390_TSCH; run 555 arch/s390/kvm/priv.c vcpu->run->s390_tsch.dequeued = !!inti; run 557 arch/s390/kvm/priv.c vcpu->run->s390_tsch.subchannel_id = inti->io.subchannel_id; run 558 arch/s390/kvm/priv.c vcpu->run->s390_tsch.subchannel_nr = inti->io.subchannel_nr; run 559 arch/s390/kvm/priv.c vcpu->run->s390_tsch.io_int_parm = inti->io.io_int_parm; run 560 arch/s390/kvm/priv.c vcpu->run->s390_tsch.io_int_word = inti->io.io_int_word; run 562 arch/s390/kvm/priv.c vcpu->run->s390_tsch.ipb = vcpu->arch.sie_block->ipb; run 632 arch/s390/kvm/priv.c reg0 = vcpu->run->s.regs.gprs[0]; run 666 arch/s390/kvm/priv.c if (!ret && vcpu->run->s.regs.gprs[1] & 0x00ff0000) run 676 arch/s390/kvm/priv.c memcpy(&vcpu->run->s.regs.gprs[1], &status, sizeof(status)); run 836 arch/s390/kvm/priv.c vcpu->run->exit_reason = KVM_EXIT_S390_STSI; run 837 arch/s390/kvm/priv.c vcpu->run->s390_stsi.addr = addr; run 838 arch/s390/kvm/priv.c vcpu->run->s390_stsi.ar = ar; run 839 arch/s390/kvm/priv.c vcpu->run->s390_stsi.fc = fc; run 840 arch/s390/kvm/priv.c vcpu->run->s390_stsi.sel1 = sel1; run 841 arch/s390/kvm/priv.c vcpu->run->s390_stsi.sel2 = sel2; run 846 arch/s390/kvm/priv.c int fc = (vcpu->run->s.regs.gprs[0] & 0xf0000000) >> 28; run 847 arch/s390/kvm/priv.c int sel1 = vcpu->run->s.regs.gprs[0] & 0xff; run 848 arch/s390/kvm/priv.c int sel2 = vcpu->run->s.regs.gprs[1] & 0xffff; run 865 arch/s390/kvm/priv.c if (vcpu->run->s.regs.gprs[0] & 0x0fffff00 run 866 arch/s390/kvm/priv.c || vcpu->run->s.regs.gprs[1] & 0xffff0000) run 870 arch/s390/kvm/priv.c vcpu->run->s.regs.gprs[0] = 3 << 28; run 911 arch/s390/kvm/priv.c vcpu->run->s.regs.gprs[0] = 0; run 987 arch/s390/kvm/priv.c vcpu->run->s.regs.gprs[reg1] &= 0xffffffff00000000UL; run 988 arch/s390/kvm/priv.c vcpu->run->s.regs.gprs[reg1] |= vcpu->arch.sie_block->gpsw.mask >> 32; run 990 arch/s390/kvm/priv.c vcpu->run->s.regs.gprs[reg2] &= 0xffffffff00000000UL; run 991 arch/s390/kvm/priv.c vcpu->run->s.regs.gprs[reg2] |= run 1024 arch/s390/kvm/priv.c if (vcpu->run->s.regs.gprs[reg1] & PFMF_RESERVED) run 1028 arch/s390/kvm/priv.c if (vcpu->run->s.regs.gprs[reg1] & PFMF_NQ && run 1033 arch/s390/kvm/priv.c if (vcpu->run->s.regs.gprs[reg1] & PFMF_SK && run 1035 arch/s390/kvm/priv.c mr = vcpu->run->s.regs.gprs[reg1] & PFMF_MR; run 1036 arch/s390/kvm/priv.c mc = vcpu->run->s.regs.gprs[reg1] & PFMF_MC; run 1039 arch/s390/kvm/priv.c nq = vcpu->run->s.regs.gprs[reg1] & PFMF_NQ; run 1040 arch/s390/kvm/priv.c key = vcpu->run->s.regs.gprs[reg1] & PFMF_KEY; run 1041 arch/s390/kvm/priv.c start = vcpu->run->s.regs.gprs[reg2] & PAGE_MASK; run 1044 arch/s390/kvm/priv.c if (vcpu->run->s.regs.gprs[reg1] & PFMF_CF) { run 1049 arch/s390/kvm/priv.c switch (vcpu->run->s.regs.gprs[reg1] & PFMF_FSC) { run 1079 arch/s390/kvm/priv.c if (vcpu->run->s.regs.gprs[reg1] & PFMF_CF) { run 1084 arch/s390/kvm/priv.c if (vcpu->run->s.regs.gprs[reg1] & PFMF_SK) { run 1107 arch/s390/kvm/priv.c if (vcpu->run->s.regs.gprs[reg1] & PFMF_FSC) { run 1109 arch/s390/kvm/priv.c vcpu->run->s.regs.gprs[reg2] = end; run 1111 arch/s390/kvm/priv.c vcpu->run->s.regs.gprs[reg2] &= ~0xffffffffUL; run 1113 arch/s390/kvm/priv.c vcpu->run->s.regs.gprs[reg2] |= end; run 1134 arch/s390/kvm/priv.c gfn = vcpu->run->s.regs.gprs[r2] >> PAGE_SHIFT; run 1144 arch/s390/kvm/priv.c vcpu->run->s.regs.gprs[r1] = res; /* Exception Indication */ run 1161 arch/s390/kvm/priv.c vcpu->run->s.regs.gprs[r1] = res; run 1509 arch/s390/kvm/priv.c if (vcpu->run->s.regs.gprs[0] & 0x00000000ffff0000) run 1513 arch/s390/kvm/priv.c value = vcpu->run->s.regs.gprs[0] & 0x000000000000ffff; run 414 arch/s390/kvm/sigp.c u16 cpu_addr = vcpu->run->s.regs.gprs[r3]; run 427 arch/s390/kvm/sigp.c parameter = vcpu->run->s.regs.gprs[r1]; run 429 arch/s390/kvm/sigp.c parameter = vcpu->run->s.regs.gprs[r1 + 1]; run 436 arch/s390/kvm/sigp.c &vcpu->run->s.regs.gprs[r1]); run 441 arch/s390/kvm/sigp.c &vcpu->run->s.regs.gprs[r1]); run 463 arch/s390/kvm/sigp.c u16 cpu_addr = vcpu->run->s.regs.gprs[r3]; run 1036 arch/s390/kvm/vsie.c rc = sie64a(scb_s, vcpu->run->s.regs.gprs); run 1076 arch/x86/include/asm/kvm_host.h void (*run)(struct kvm_vcpu *vcpu); run 1556 arch/x86/kvm/hyperv.c return kvm_hv_hypercall_complete(vcpu, vcpu->run->hyperv.u.hcall.result); run 1660 arch/x86/kvm/hyperv.c vcpu->run->exit_reason = KVM_EXIT_HYPERV; run 1661 arch/x86/kvm/hyperv.c vcpu->run->hyperv.type = KVM_EXIT_HYPERV_HCALL; run 1662 arch/x86/kvm/hyperv.c vcpu->run->hyperv.u.hcall.input = param; run 1663 arch/x86/kvm/hyperv.c vcpu->run->hyperv.u.hcall.params[0] = ingpa; run 1664 arch/x86/kvm/hyperv.c vcpu->run->hyperv.u.hcall.params[1] = outgpa; run 1247 arch/x86/kvm/lapic.c struct kvm_run *run = vcpu->run; run 1250 arch/x86/kvm/lapic.c run->tpr_access.rip = kvm_rip_read(vcpu); run 1251 arch/x86/kvm/lapic.c run->tpr_access.is_write = write; run 2772 arch/x86/kvm/svm.c struct kvm_run *kvm_run = svm->vcpu.run; run 2802 arch/x86/kvm/svm.c struct kvm_run *kvm_run = svm->vcpu.run; run 2910 arch/x86/kvm/svm.c struct kvm_run *kvm_run = svm->vcpu.run; run 4139 arch/x86/kvm/svm.c struct kvm_run *kvm_run = svm->vcpu.run; run 4979 arch/x86/kvm/svm.c struct kvm_run *kvm_run = vcpu->run; run 5038 arch/x86/kvm/svm.c vcpu->run->exit_reason = KVM_EXIT_INTERNAL_ERROR; run 5039 arch/x86/kvm/svm.c vcpu->run->internal.suberror = run 5041 arch/x86/kvm/svm.c vcpu->run->internal.ndata = 1; run 5042 arch/x86/kvm/svm.c vcpu->run->internal.data[0] = exit_code; run 7294 arch/x86/kvm/svm.c .run = svm_vcpu_run, run 2958 arch/x86/kvm/vmx/nested.c vcpu->run->exit_reason = KVM_EXIT_INTERNAL_ERROR; run 2959 arch/x86/kvm/vmx/nested.c vcpu->run->internal.suberror = run 2961 arch/x86/kvm/vmx/nested.c vcpu->run->internal.ndata = 0; run 4597 arch/x86/kvm/vmx/vmx.c struct kvm_run *kvm_run = vcpu->run; run 4637 arch/x86/kvm/vmx/vmx.c vcpu->run->exit_reason = KVM_EXIT_INTERNAL_ERROR; run 4638 arch/x86/kvm/vmx/vmx.c vcpu->run->internal.suberror = KVM_INTERNAL_ERROR_SIMUL_EX; run 4639 arch/x86/kvm/vmx/vmx.c vcpu->run->internal.ndata = 3; run 4640 arch/x86/kvm/vmx/vmx.c vcpu->run->internal.data[0] = vect_info; run 4641 arch/x86/kvm/vmx/vmx.c vcpu->run->internal.data[1] = intr_info; run 4642 arch/x86/kvm/vmx/vmx.c vcpu->run->internal.data[2] = error_code; run 4707 arch/x86/kvm/vmx/vmx.c vcpu->run->exit_reason = KVM_EXIT_SHUTDOWN; run 4841 arch/x86/kvm/vmx/vmx.c vcpu->run->exit_reason = KVM_EXIT_SET_TPR; run 4875 arch/x86/kvm/vmx/vmx.c vcpu->run->exit_reason = 0; run 4904 arch/x86/kvm/vmx/vmx.c vcpu->run->debug.arch.dr6 = vcpu->arch.dr6; run 4905 arch/x86/kvm/vmx/vmx.c vcpu->run->debug.arch.dr7 = dr7; run 4906 arch/x86/kvm/vmx/vmx.c vcpu->run->debug.arch.pc = kvm_get_linear_rip(vcpu); run 4907 arch/x86/kvm/vmx/vmx.c vcpu->run->debug.arch.exception = DB_VECTOR; run 4908 arch/x86/kvm/vmx/vmx.c vcpu->run->exit_reason = KVM_EXIT_DEBUG; run 5247 arch/x86/kvm/vmx/vmx.c vcpu->run->exit_reason = KVM_EXIT_INTERNAL_ERROR; run 5248 arch/x86/kvm/vmx/vmx.c vcpu->run->internal.suberror = run 5250 arch/x86/kvm/vmx/vmx.c vcpu->run->internal.ndata = 0; run 5871 arch/x86/kvm/vmx/vmx.c vcpu->run->exit_reason = KVM_EXIT_FAIL_ENTRY; run 5872 arch/x86/kvm/vmx/vmx.c vcpu->run->fail_entry.hardware_entry_failure_reason run 5879 arch/x86/kvm/vmx/vmx.c vcpu->run->exit_reason = KVM_EXIT_FAIL_ENTRY; run 5880 arch/x86/kvm/vmx/vmx.c vcpu->run->fail_entry.hardware_entry_failure_reason run 5897 arch/x86/kvm/vmx/vmx.c vcpu->run->exit_reason = KVM_EXIT_INTERNAL_ERROR; run 5898 arch/x86/kvm/vmx/vmx.c vcpu->run->internal.suberror = KVM_INTERNAL_ERROR_DELIVERY_EV; run 5899 arch/x86/kvm/vmx/vmx.c vcpu->run->internal.ndata = 3; run 5900 arch/x86/kvm/vmx/vmx.c vcpu->run->internal.data[0] = vectoring_info; run 5901 arch/x86/kvm/vmx/vmx.c vcpu->run->internal.data[1] = exit_reason; run 5902 arch/x86/kvm/vmx/vmx.c vcpu->run->internal.data[2] = vcpu->arch.exit_qualification; run 5904 arch/x86/kvm/vmx/vmx.c vcpu->run->internal.ndata++; run 5905 arch/x86/kvm/vmx/vmx.c vcpu->run->internal.data[3] = run 5937 arch/x86/kvm/vmx/vmx.c vcpu->run->exit_reason = KVM_EXIT_INTERNAL_ERROR; run 5938 arch/x86/kvm/vmx/vmx.c vcpu->run->internal.suberror = run 5940 arch/x86/kvm/vmx/vmx.c vcpu->run->internal.ndata = 1; run 5941 arch/x86/kvm/vmx/vmx.c vcpu->run->internal.data[0] = exit_reason; run 7839 arch/x86/kvm/vmx/vmx.c .run = vmx_vcpu_run, run 5675 arch/x86/kvm/x86.c memcpy(vcpu->run->mmio.data, frag->data, min(8u, frag->len)); run 5791 arch/x86/kvm/x86.c vcpu->run->mmio.len = min(8u, vcpu->mmio_fragments[0].len); run 5792 arch/x86/kvm/x86.c vcpu->run->mmio.is_write = vcpu->mmio_is_write = ops->write; run 5793 arch/x86/kvm/x86.c vcpu->run->exit_reason = KVM_EXIT_MMIO; run 5794 arch/x86/kvm/x86.c vcpu->run->mmio.phys_addr = gpa; run 5925 arch/x86/kvm/x86.c vcpu->run->exit_reason = KVM_EXIT_IO; run 5926 arch/x86/kvm/x86.c vcpu->run->io.direction = in ? KVM_EXIT_IO_IN : KVM_EXIT_IO_OUT; run 5927 arch/x86/kvm/x86.c vcpu->run->io.size = size; run 5928 arch/x86/kvm/x86.c vcpu->run->io.data_offset = KVM_PIO_PAGE_OFFSET * PAGE_SIZE; run 5929 arch/x86/kvm/x86.c vcpu->run->io.count = count; run 5930 arch/x86/kvm/x86.c vcpu->run->io.port = port; run 6416 arch/x86/kvm/x86.c vcpu->run->exit_reason = KVM_EXIT_INTERNAL_ERROR; run 6417 arch/x86/kvm/x86.c vcpu->run->internal.suberror = KVM_INTERNAL_ERROR_EMULATION; run 6418 arch/x86/kvm/x86.c vcpu->run->internal.ndata = 0; run 6425 arch/x86/kvm/x86.c vcpu->run->exit_reason = KVM_EXIT_INTERNAL_ERROR; run 6426 arch/x86/kvm/x86.c vcpu->run->internal.suberror = KVM_INTERNAL_ERROR_EMULATION; run 6427 arch/x86/kvm/x86.c vcpu->run->internal.ndata = 0; run 6588 arch/x86/kvm/x86.c struct kvm_run *kvm_run = vcpu->run; run 6628 arch/x86/kvm/x86.c struct kvm_run *kvm_run = vcpu->run; run 7354 arch/x86/kvm/x86.c vcpu->run->exit_reason = KVM_EXIT_HLT; run 7529 arch/x86/kvm/x86.c return vcpu->run->request_interrupt_window && run 7535 arch/x86/kvm/x86.c struct kvm_run *kvm_run = vcpu->run; run 8065 arch/x86/kvm/x86.c vcpu->run->exit_reason = KVM_EXIT_TPR_ACCESS; run 8070 arch/x86/kvm/x86.c vcpu->run->exit_reason = KVM_EXIT_SHUTDOWN; run 8095 arch/x86/kvm/x86.c vcpu->run->exit_reason = KVM_EXIT_IOAPIC_EOI; run 8096 arch/x86/kvm/x86.c vcpu->run->eoi.vector = run 8109 arch/x86/kvm/x86.c vcpu->run->exit_reason = KVM_EXIT_SYSTEM_EVENT; run 8110 arch/x86/kvm/x86.c vcpu->run->system_event.type = KVM_SYSTEM_EVENT_CRASH; run 8115 arch/x86/kvm/x86.c vcpu->run->exit_reason = KVM_EXIT_SYSTEM_EVENT; run 8116 arch/x86/kvm/x86.c vcpu->run->system_event.type = KVM_SYSTEM_EVENT_RESET; run 8121 arch/x86/kvm/x86.c vcpu->run->exit_reason = KVM_EXIT_HYPERV; run 8122 arch/x86/kvm/x86.c vcpu->run->hyperv = vcpu->arch.hyperv.exit; run 8253 arch/x86/kvm/x86.c kvm_x86_ops->run(vcpu); run 8408 arch/x86/kvm/x86.c vcpu->run->exit_reason = KVM_EXIT_IRQ_WINDOW_OPEN; run 8417 arch/x86/kvm/x86.c vcpu->run->exit_reason = KVM_EXIT_INTR; run 8470 arch/x86/kvm/x86.c struct kvm_run *run = vcpu->run; run 8480 arch/x86/kvm/x86.c memcpy(frag->data, run->mmio.data, len); run 8503 arch/x86/kvm/x86.c run->exit_reason = KVM_EXIT_MMIO; run 8504 arch/x86/kvm/x86.c run->mmio.phys_addr = frag->gpa; run 8506 arch/x86/kvm/x86.c memcpy(run->mmio.data, frag->data, min(8u, frag->len)); run 8507 arch/x86/kvm/x86.c run->mmio.len = min(8u, frag->len); run 8508 arch/x86/kvm/x86.c run->mmio.is_write = vcpu->mmio_is_write; run 8578 arch/x86/kvm/x86.c vcpu->run->exit_reason = KVM_EXIT_INTR; run 8584 arch/x86/kvm/x86.c if (vcpu->run->kvm_valid_regs & ~KVM_SYNC_X86_VALID_FIELDS) { run 8589 arch/x86/kvm/x86.c if (vcpu->run->kvm_dirty_regs) { run 8619 arch/x86/kvm/x86.c if (vcpu->run->kvm_valid_regs) run 8828 arch/x86/kvm/x86.c vcpu->run->exit_reason = KVM_EXIT_INTERNAL_ERROR; run 8829 arch/x86/kvm/x86.c vcpu->run->internal.suberror = KVM_INTERNAL_ERROR_EMULATION; run 8830 arch/x86/kvm/x86.c vcpu->run->internal.ndata = 0; run 9089 arch/x86/kvm/x86.c if (vcpu->run->kvm_valid_regs & KVM_SYNC_X86_REGS) run 9090 arch/x86/kvm/x86.c __get_regs(vcpu, &vcpu->run->s.regs.regs); run 9092 arch/x86/kvm/x86.c if (vcpu->run->kvm_valid_regs & KVM_SYNC_X86_SREGS) run 9093 arch/x86/kvm/x86.c __get_sregs(vcpu, &vcpu->run->s.regs.sregs); run 9095 arch/x86/kvm/x86.c if (vcpu->run->kvm_valid_regs & KVM_SYNC_X86_EVENTS) run 9097 arch/x86/kvm/x86.c vcpu, &vcpu->run->s.regs.events); run 9102 arch/x86/kvm/x86.c if (vcpu->run->kvm_dirty_regs & ~KVM_SYNC_X86_VALID_FIELDS) run 9105 arch/x86/kvm/x86.c if (vcpu->run->kvm_dirty_regs & KVM_SYNC_X86_REGS) { run 9106 arch/x86/kvm/x86.c __set_regs(vcpu, &vcpu->run->s.regs.regs); run 9107 arch/x86/kvm/x86.c vcpu->run->kvm_dirty_regs &= ~KVM_SYNC_X86_REGS; run 9109 arch/x86/kvm/x86.c if (vcpu->run->kvm_dirty_regs & KVM_SYNC_X86_SREGS) { run 9110 arch/x86/kvm/x86.c if (__set_sregs(vcpu, &vcpu->run->s.regs.sregs)) run 9112 arch/x86/kvm/x86.c vcpu->run->kvm_dirty_regs &= ~KVM_SYNC_X86_SREGS; run 9114 arch/x86/kvm/x86.c if (vcpu->run->kvm_dirty_regs & KVM_SYNC_X86_EVENTS) { run 9116 arch/x86/kvm/x86.c vcpu, &vcpu->run->s.regs.events)) run 9118 arch/x86/kvm/x86.c vcpu->run->kvm_dirty_regs &= ~KVM_SYNC_X86_EVENTS; run 596 block/blk-mq-debugfs.c seq_printf(m, "%lu\n", hctx->run); run 605 block/blk-mq-debugfs.c hctx->run = 0; run 181 block/blk-mq-sched.c hctx->run++; run 395 block/blk-mq-sched.c goto run; run 424 block/blk-mq-sched.c goto run; run 438 block/blk-mq-sched.c run: run 154 drivers/acpi/apei/apei-base.c apei_exec_ins_func_t run; run 172 drivers/acpi/apei/apei-base.c !ctx->ins_table[entry->instruction].run) { run 178 drivers/acpi/apei/apei-base.c run = ctx->ins_table[entry->instruction].run; run 179 drivers/acpi/apei/apei-base.c rc = run(ctx, entry); run 213 drivers/acpi/apei/apei-base.c if (ins >= ctx->instructions || !ins_table[ins].run) { run 22 drivers/acpi/apei/apei-internal.h apei_exec_ins_func_t run; run 111 drivers/acpi/apei/einj.c .run = apei_exec_read_register, run 115 drivers/acpi/apei/einj.c .run = apei_exec_read_register_value, run 119 drivers/acpi/apei/einj.c .run = apei_exec_write_register, run 123 drivers/acpi/apei/einj.c .run = apei_exec_write_register_value, run 127 drivers/acpi/apei/einj.c .run = apei_exec_noop, run 292 drivers/acpi/apei/erst.c .run = apei_exec_read_register, run 296 drivers/acpi/apei/erst.c .run = apei_exec_read_register_value, run 300 drivers/acpi/apei/erst.c .run = apei_exec_write_register, run 304 drivers/acpi/apei/erst.c .run = apei_exec_write_register_value, run 308 drivers/acpi/apei/erst.c .run = apei_exec_noop, run 312 drivers/acpi/apei/erst.c .run = erst_exec_load_var1, run 316 drivers/acpi/apei/erst.c .run = erst_exec_load_var2, run 320 drivers/acpi/apei/erst.c .run = erst_exec_store_var1, run 324 drivers/acpi/apei/erst.c .run = erst_exec_add, run 328 drivers/acpi/apei/erst.c .run = erst_exec_subtract, run 332 drivers/acpi/apei/erst.c .run = erst_exec_add_value, run 336 drivers/acpi/apei/erst.c .run = erst_exec_subtract_value, run 340 drivers/acpi/apei/erst.c .run = erst_exec_stall, run 344 drivers/acpi/apei/erst.c .run = erst_exec_stall_while_true, run 348 drivers/acpi/apei/erst.c .run = erst_exec_skip_next_instruction_if_true, run 352 drivers/acpi/apei/erst.c .run = erst_exec_goto, run 356 drivers/acpi/apei/erst.c .run = erst_exec_set_src_address_base, run 360 drivers/acpi/apei/erst.c .run = erst_exec_set_dst_address_base, run 364 drivers/acpi/apei/erst.c .run = erst_exec_move_data, run 146 drivers/dma/dmatest.c module_param_cb(run, &run_ops, &dmatest_run, S_IRUGO | S_IWUSR); run 147 drivers/dma/dmatest.c MODULE_PARM_DESC(run, "Run the test (default: false)"); run 93 drivers/gpu/drm/tiny/gm12u320.c bool run; run 354 drivers/gpu/drm/tiny/gm12u320.c while (gm12u320->fb_update.run) { run 418 drivers/gpu/drm/tiny/gm12u320.c !gm12u320->fb_update.run || run 465 drivers/gpu/drm/tiny/gm12u320.c gm12u320->fb_update.run = true; run 474 drivers/gpu/drm/tiny/gm12u320.c gm12u320->fb_update.run = false; run 1241 drivers/gpu/ipu-v3/ipu-image-convert.c struct ipu_image_convert_run *run; run 1246 drivers/gpu/ipu-v3/ipu-image-convert.c list_for_each_entry(run, q, list) { run 1247 drivers/gpu/ipu-v3/ipu-image-convert.c if (run->ctx == ctx) run 1254 drivers/gpu/ipu-v3/ipu-image-convert.c static void convert_stop(struct ipu_image_convert_run *run) run 1256 drivers/gpu/ipu-v3/ipu-image-convert.c struct ipu_image_convert_ctx *ctx = run->ctx; run 1261 drivers/gpu/ipu-v3/ipu-image-convert.c __func__, chan->ic_task, ctx, run); run 1367 drivers/gpu/ipu-v3/ipu-image-convert.c static int convert_start(struct ipu_image_convert_run *run, unsigned int tile) run 1369 drivers/gpu/ipu-v3/ipu-image-convert.c struct ipu_image_convert_ctx *ctx = run->ctx; run 1381 drivers/gpu/ipu-v3/ipu-image-convert.c __func__, chan->ic_task, ctx, run, tile, dst_tile); run 1479 drivers/gpu/ipu-v3/ipu-image-convert.c static int do_run(struct ipu_image_convert_run *run) run 1481 drivers/gpu/ipu-v3/ipu-image-convert.c struct ipu_image_convert_ctx *ctx = run->ctx; run 1486 drivers/gpu/ipu-v3/ipu-image-convert.c ctx->in.base.phys0 = run->in_phys; run 1487 drivers/gpu/ipu-v3/ipu-image-convert.c ctx->out.base.phys0 = run->out_phys; run 1493 drivers/gpu/ipu-v3/ipu-image-convert.c list_del(&run->list); run 1494 drivers/gpu/ipu-v3/ipu-image-convert.c chan->current_run = run; run 1496 drivers/gpu/ipu-v3/ipu-image-convert.c return convert_start(run, 0); run 1503 drivers/gpu/ipu-v3/ipu-image-convert.c struct ipu_image_convert_run *run, *tmp; run 1508 drivers/gpu/ipu-v3/ipu-image-convert.c list_for_each_entry_safe(run, tmp, &chan->pending_q, list) { run 1510 drivers/gpu/ipu-v3/ipu-image-convert.c if (run->ctx->aborting) { run 1513 drivers/gpu/ipu-v3/ipu-image-convert.c __func__, chan->ic_task, run->ctx, run); run 1517 drivers/gpu/ipu-v3/ipu-image-convert.c ret = do_run(run); run 1526 drivers/gpu/ipu-v3/ipu-image-convert.c run->status = ret; run 1527 drivers/gpu/ipu-v3/ipu-image-convert.c list_add_tail(&run->list, &chan->done_q); run 1535 drivers/gpu/ipu-v3/ipu-image-convert.c struct ipu_image_convert_run *run; run 1541 drivers/gpu/ipu-v3/ipu-image-convert.c run = list_entry(chan->done_q.next, run 1545 drivers/gpu/ipu-v3/ipu-image-convert.c list_del(&run->list); run 1549 drivers/gpu/ipu-v3/ipu-image-convert.c __func__, chan->ic_task, run->ctx, run, run->status); run 1553 drivers/gpu/ipu-v3/ipu-image-convert.c run->ctx->complete(run, run->ctx->complete_context); run 1618 drivers/gpu/ipu-v3/ipu-image-convert.c static irqreturn_t do_irq(struct ipu_image_convert_run *run) run 1620 drivers/gpu/ipu-v3/ipu-image-convert.c struct ipu_image_convert_ctx *ctx = run->ctx; run 1642 drivers/gpu/ipu-v3/ipu-image-convert.c convert_stop(run); run 1643 drivers/gpu/ipu-v3/ipu-image-convert.c run->status = -EIO; run 1651 drivers/gpu/ipu-v3/ipu-image-convert.c convert_stop(run); run 1652 drivers/gpu/ipu-v3/ipu-image-convert.c run->status = 0; run 1661 drivers/gpu/ipu-v3/ipu-image-convert.c convert_stop(run); run 1662 drivers/gpu/ipu-v3/ipu-image-convert.c convert_start(run, ctx->next_tile); run 1706 drivers/gpu/ipu-v3/ipu-image-convert.c list_add_tail(&run->list, &chan->done_q); run 1716 drivers/gpu/ipu-v3/ipu-image-convert.c struct ipu_image_convert_run *run; run 1723 drivers/gpu/ipu-v3/ipu-image-convert.c run = chan->current_run; run 1724 drivers/gpu/ipu-v3/ipu-image-convert.c if (!run) { run 1729 drivers/gpu/ipu-v3/ipu-image-convert.c ctx = run->ctx; run 1737 drivers/gpu/ipu-v3/ipu-image-convert.c ret = do_irq(run); run 1748 drivers/gpu/ipu-v3/ipu-image-convert.c struct ipu_image_convert_run *run; run 1755 drivers/gpu/ipu-v3/ipu-image-convert.c run = chan->current_run; run 1756 drivers/gpu/ipu-v3/ipu-image-convert.c if (!run) { run 1761 drivers/gpu/ipu-v3/ipu-image-convert.c ctx = run->ctx; run 1770 drivers/gpu/ipu-v3/ipu-image-convert.c ret = do_irq(run); run 1783 drivers/gpu/ipu-v3/ipu-image-convert.c struct ipu_image_convert_run *run; run 1788 drivers/gpu/ipu-v3/ipu-image-convert.c run = chan->current_run; run 1789 drivers/gpu/ipu-v3/ipu-image-convert.c if (run && run->ctx == ctx) { run 1790 drivers/gpu/ipu-v3/ipu-image-convert.c convert_stop(run); run 1791 drivers/gpu/ipu-v3/ipu-image-convert.c run->status = -EIO; run 1792 drivers/gpu/ipu-v3/ipu-image-convert.c list_add_tail(&run->list, &chan->done_q); run 2234 drivers/gpu/ipu-v3/ipu-image-convert.c int ipu_image_convert_queue(struct ipu_image_convert_run *run) run 2242 drivers/gpu/ipu-v3/ipu-image-convert.c if (!run || !run->ctx || !run->in_phys || !run->out_phys) run 2245 drivers/gpu/ipu-v3/ipu-image-convert.c ctx = run->ctx; run 2250 drivers/gpu/ipu-v3/ipu-image-convert.c chan->ic_task, ctx, run); run 2252 drivers/gpu/ipu-v3/ipu-image-convert.c INIT_LIST_HEAD(&run->list); run 2261 drivers/gpu/ipu-v3/ipu-image-convert.c list_add_tail(&run->list, &chan->pending_q); run 2264 drivers/gpu/ipu-v3/ipu-image-convert.c ret = do_run(run); run 2279 drivers/gpu/ipu-v3/ipu-image-convert.c struct ipu_image_convert_run *run, *active_run, *tmp; run 2286 drivers/gpu/ipu-v3/ipu-image-convert.c list_for_each_entry_safe(run, tmp, &chan->pending_q, list) { run 2287 drivers/gpu/ipu-v3/ipu-image-convert.c if (run->ctx != ctx) run 2289 drivers/gpu/ipu-v3/ipu-image-convert.c run->status = -EIO; run 2290 drivers/gpu/ipu-v3/ipu-image-convert.c list_move_tail(&run->list, &chan->done_q); run 2380 drivers/gpu/ipu-v3/ipu-image-convert.c struct ipu_image_convert_run *run; run 2388 drivers/gpu/ipu-v3/ipu-image-convert.c run = kzalloc(sizeof(*run), GFP_KERNEL); run 2389 drivers/gpu/ipu-v3/ipu-image-convert.c if (!run) { run 2394 drivers/gpu/ipu-v3/ipu-image-convert.c run->ctx = ctx; run 2395 drivers/gpu/ipu-v3/ipu-image-convert.c run->in_phys = in->phys0; run 2396 drivers/gpu/ipu-v3/ipu-image-convert.c run->out_phys = out->phys0; run 2398 drivers/gpu/ipu-v3/ipu-image-convert.c ret = ipu_image_convert_queue(run); run 2401 drivers/gpu/ipu-v3/ipu-image-convert.c kfree(run); run 2405 drivers/gpu/ipu-v3/ipu-image-convert.c return run; run 2410 drivers/gpu/ipu-v3/ipu-image-convert.c static void image_convert_sync_complete(struct ipu_image_convert_run *run, run 2422 drivers/gpu/ipu-v3/ipu-image-convert.c struct ipu_image_convert_run *run; run 2428 drivers/gpu/ipu-v3/ipu-image-convert.c run = ipu_image_convert(ipu, ic_task, in, out, rot_mode, run 2430 drivers/gpu/ipu-v3/ipu-image-convert.c if (IS_ERR(run)) run 2431 drivers/gpu/ipu-v3/ipu-image-convert.c return PTR_ERR(run); run 2436 drivers/gpu/ipu-v3/ipu-image-convert.c ipu_image_convert_unprepare(run->ctx); run 2437 drivers/gpu/ipu-v3/ipu-image-convert.c kfree(run); run 469 drivers/iio/light/si1145.c int run = 1; run 471 drivers/iio/light/si1145.c while (i + run < indio_dev->masklength) { run 472 drivers/iio/light/si1145.c if (!test_bit(i + run, indio_dev->active_scan_mask)) run 474 drivers/iio/light/si1145.c if (indio_dev->channels[i + run].address != run 475 drivers/iio/light/si1145.c indio_dev->channels[i].address + 2 * run) run 477 drivers/iio/light/si1145.c run++; run 482 drivers/iio/light/si1145.c sizeof(u16) * run, &buffer[j]); run 485 drivers/iio/light/si1145.c j += run * sizeof(u16); run 486 drivers/iio/light/si1145.c i += run - 1; run 340 drivers/md/md-faulty.c .run = faulty_run, run 319 drivers/md/md-linear.c .run = linear_run, run 474 drivers/md/md-multipath.c .run = multipath_run, run 3976 drivers/md/md.c pers->run(mddev); run 5743 drivers/md/md.c err = pers->run(mddev); run 7710 drivers/md/md.c thread->run(thread); run 7726 drivers/md/md.c struct md_thread *md_register_thread(void (*run) (struct md_thread *), run 7737 drivers/md/md.c thread->run = run; run 558 drivers/md/md.h int (*run)(struct mddev *mddev); run 666 drivers/md/md.h void (*run) (struct md_thread *thread); run 690 drivers/md/md.h void (*run)(struct md_thread *thread), run 815 drivers/md/raid0.c .run = raid0_run, run 3373 drivers/md/raid1.c .run = raid1_run, run 4916 drivers/md/raid10.c .run = raid10_run, run 8415 drivers/md/raid5.c .run = raid5_run, run 8440 drivers/md/raid5.c .run = raid5_run, run 8466 drivers/md/raid5.c .run = raid5_run, run 203 drivers/media/platform/mtk-vpu/mtk_vpu.c struct vpu_run run; run 419 drivers/media/platform/mtk-vpu/mtk_vpu.c return vpu->run.dec_capability; run 427 drivers/media/platform/mtk-vpu/mtk_vpu.c return vpu->run.enc_capability; run 533 drivers/media/platform/mtk-vpu/mtk_vpu.c struct vpu_run *run; run 542 drivers/media/platform/mtk-vpu/mtk_vpu.c run = &vpu->run; run 559 drivers/media/platform/mtk-vpu/mtk_vpu.c run->signaled = false; run 579 drivers/media/platform/mtk-vpu/mtk_vpu.c ret = wait_event_interruptible_timeout(run->wq, run 580 drivers/media/platform/mtk-vpu/mtk_vpu.c run->signaled, run 593 drivers/media/platform/mtk-vpu/mtk_vpu.c dev_info(dev, "vpu is ready. Fw version %s\n", run->fw_ver); run 606 drivers/media/platform/mtk-vpu/mtk_vpu.c struct vpu_run *run = (struct vpu_run *)data; run 608 drivers/media/platform/mtk-vpu/mtk_vpu.c vpu->run.signaled = run->signaled; run 609 drivers/media/platform/mtk-vpu/mtk_vpu.c strscpy(vpu->run.fw_ver, run->fw_ver, sizeof(vpu->run.fw_ver)); run 610 drivers/media/platform/mtk-vpu/mtk_vpu.c vpu->run.dec_capability = run->dec_capability; run 611 drivers/media/platform/mtk-vpu/mtk_vpu.c vpu->run.enc_capability = run->enc_capability; run 612 drivers/media/platform/mtk-vpu/mtk_vpu.c wake_up_interruptible(&vpu->run.wq); run 647 drivers/media/platform/mtk-vpu/mtk_vpu.c vpu->run.fw_ver, pc, wdt, run 872 drivers/media/platform/mtk-vpu/mtk_vpu.c init_waitqueue_head(&vpu->run.wq); run 1020 drivers/mmc/core/mmc_test.c int (*run)(struct mmc_test_card *); run 2551 drivers/mmc/core/mmc_test.c .run = mmc_test_basic_write, run 2556 drivers/mmc/core/mmc_test.c .run = mmc_test_basic_read, run 2562 drivers/mmc/core/mmc_test.c .run = mmc_test_verify_write, run 2569 drivers/mmc/core/mmc_test.c .run = mmc_test_verify_read, run 2576 drivers/mmc/core/mmc_test.c .run = mmc_test_multi_write, run 2583 drivers/mmc/core/mmc_test.c .run = mmc_test_multi_read, run 2590 drivers/mmc/core/mmc_test.c .run = mmc_test_pow2_write, run 2597 drivers/mmc/core/mmc_test.c .run = mmc_test_pow2_read, run 2604 drivers/mmc/core/mmc_test.c .run = mmc_test_weird_write, run 2611 drivers/mmc/core/mmc_test.c .run = mmc_test_weird_read, run 2618 drivers/mmc/core/mmc_test.c .run = mmc_test_align_write, run 2625 drivers/mmc/core/mmc_test.c .run = mmc_test_align_read, run 2632 drivers/mmc/core/mmc_test.c .run = mmc_test_align_multi_write, run 2639 drivers/mmc/core/mmc_test.c .run = mmc_test_align_multi_read, run 2645 drivers/mmc/core/mmc_test.c .run = mmc_test_xfersize_write, run 2650 drivers/mmc/core/mmc_test.c .run = mmc_test_xfersize_read, run 2655 drivers/mmc/core/mmc_test.c .run = mmc_test_multi_xfersize_write, run 2660 drivers/mmc/core/mmc_test.c .run = mmc_test_multi_xfersize_read, run 2668 drivers/mmc/core/mmc_test.c .run = mmc_test_write_high, run 2675 drivers/mmc/core/mmc_test.c .run = mmc_test_read_high, run 2682 drivers/mmc/core/mmc_test.c .run = mmc_test_multi_write_high, run 2689 drivers/mmc/core/mmc_test.c .run = mmc_test_multi_read_high, run 2697 drivers/mmc/core/mmc_test.c .run = mmc_test_no_highmem, run 2702 drivers/mmc/core/mmc_test.c .run = mmc_test_no_highmem, run 2707 drivers/mmc/core/mmc_test.c .run = mmc_test_no_highmem, run 2712 drivers/mmc/core/mmc_test.c .run = mmc_test_no_highmem, run 2720 drivers/mmc/core/mmc_test.c .run = mmc_test_best_read_performance, run 2727 drivers/mmc/core/mmc_test.c .run = mmc_test_best_write_performance, run 2734 drivers/mmc/core/mmc_test.c .run = mmc_test_best_read_perf_max_scatter, run 2741 drivers/mmc/core/mmc_test.c .run = mmc_test_best_write_perf_max_scatter, run 2748 drivers/mmc/core/mmc_test.c .run = mmc_test_profile_read_perf, run 2755 drivers/mmc/core/mmc_test.c .run = mmc_test_profile_write_perf, run 2762 drivers/mmc/core/mmc_test.c .run = mmc_test_profile_trim_perf, run 2769 drivers/mmc/core/mmc_test.c .run = mmc_test_profile_seq_read_perf, run 2776 drivers/mmc/core/mmc_test.c .run = mmc_test_profile_seq_write_perf, run 2783 drivers/mmc/core/mmc_test.c .run = mmc_test_profile_seq_trim_perf, run 2790 drivers/mmc/core/mmc_test.c .run = mmc_test_random_read_perf, run 2797 drivers/mmc/core/mmc_test.c .run = mmc_test_random_write_perf, run 2804 drivers/mmc/core/mmc_test.c .run = mmc_test_large_seq_read_perf, run 2811 drivers/mmc/core/mmc_test.c .run = mmc_test_large_seq_write_perf, run 2818 drivers/mmc/core/mmc_test.c .run = mmc_test_profile_mult_write_blocking_perf, run 2825 drivers/mmc/core/mmc_test.c .run = mmc_test_profile_mult_write_nonblock_perf, run 2832 drivers/mmc/core/mmc_test.c .run = mmc_test_profile_mult_read_blocking_perf, run 2839 drivers/mmc/core/mmc_test.c .run = mmc_test_profile_mult_read_nonblock_perf, run 2846 drivers/mmc/core/mmc_test.c .run = mmc_test_profile_sglen_wr_blocking_perf, run 2853 drivers/mmc/core/mmc_test.c .run = mmc_test_profile_sglen_wr_nonblock_perf, run 2860 drivers/mmc/core/mmc_test.c .run = mmc_test_profile_sglen_r_blocking_perf, run 2867 drivers/mmc/core/mmc_test.c .run = mmc_test_profile_sglen_r_nonblock_perf, run 2873 drivers/mmc/core/mmc_test.c .run = mmc_test_reset, run 2879 drivers/mmc/core/mmc_test.c .run = mmc_test_cmds_during_read, run 2886 drivers/mmc/core/mmc_test.c .run = mmc_test_cmds_during_write, run 2893 drivers/mmc/core/mmc_test.c .run = mmc_test_cmds_during_read_cmd23, run 2900 drivers/mmc/core/mmc_test.c .run = mmc_test_cmds_during_write_cmd23, run 2907 drivers/mmc/core/mmc_test.c .run = mmc_test_cmds_during_read_cmd23_nonblock, run 2914 drivers/mmc/core/mmc_test.c .run = mmc_test_cmds_during_write_cmd23_nonblock, run 2970 drivers/mmc/core/mmc_test.c ret = mmc_test_cases[i].run(test); run 359 drivers/net/phy/phy_device.c int (*run)(struct phy_device *)) run 369 drivers/net/phy/phy_device.c fixup->run = run; run 381 drivers/net/phy/phy_device.c int (*run)(struct phy_device *)) run 383 drivers/net/phy/phy_device.c return phy_register_fixup(PHY_ANY_ID, phy_uid, phy_uid_mask, run); run 389 drivers/net/phy/phy_device.c int (*run)(struct phy_device *)) run 391 drivers/net/phy/phy_device.c return phy_register_fixup(bus_id, PHY_ANY_UID, 0xffffffff, run); run 467 drivers/net/phy/phy_device.c int err = fixup->run(phydev); run 428 drivers/net/wan/farsync.c int run; /* Port is running */ run 987 drivers/net/wan/farsync.c if (port->run) run 997 drivers/net/wan/farsync.c if (port->run) run 1336 drivers/net/wan/farsync.c if (!port->run) run 1435 drivers/net/wan/farsync.c if (!port->run) run 1562 drivers/net/wan/farsync.c if (port->run) run 1570 drivers/net/wan/farsync.c if (port->run) run 2138 drivers/net/wan/farsync.c if (port->run) { run 2142 drivers/net/wan/farsync.c port->run = 0; run 2150 drivers/net/wan/farsync.c port->run = 1; run 2169 drivers/net/wan/farsync.c if (port->run) { run 2170 drivers/net/wan/farsync.c port->run = 0; run 2500 drivers/net/wan/farsync.c card->ports[i].run = 0; run 1402 drivers/net/wireless/ath/carl9170/tx.c bool run = false; run 1450 drivers/net/wireless/ath/carl9170/tx.c run = true; run 1457 drivers/net/wireless/ath/carl9170/tx.c return run; run 1478 drivers/net/wireless/ath/carl9170/tx.c bool run; run 1508 drivers/net/wireless/ath/carl9170/tx.c run = carl9170_tx_ampdu_queue(ar, sta, skb, info); run 1509 drivers/net/wireless/ath/carl9170/tx.c if (run) run 1078 drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c err = escan->run(cfg, ifp, request); run 1126 drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c cfg->escan_info.run = brcmf_run_escan; run 3300 drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c cfg->escan_info.run = brcmf_run_escan; run 217 drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.h s32 (*run)(struct brcmf_cfg80211_info *cfg, struct brcmf_if *ifp, run 896 drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c cfg->escan_info.run = brcmf_p2p_run_escan; run 288 drivers/s390/crypto/vfio_ap_ops.c apqn = vcpu->run->s.regs.gprs[0] & 0xffff; run 300 drivers/s390/crypto/vfio_ap_ops.c status = vcpu->run->s.regs.gprs[1]; run 305 drivers/s390/crypto/vfio_ap_ops.c vcpu->run->s.regs.gprs[2]); run 310 drivers/s390/crypto/vfio_ap_ops.c memcpy(&vcpu->run->s.regs.gprs[1], &qstatus, sizeof(qstatus)); run 311 drivers/s390/crypto/vfio_ap_ops.c vcpu->run->s.regs.gprs[1] >>= 32; run 357 drivers/s390/scsi/zfcp_dbf.c rec->u.run.fsf_req_id = erp->fsf_req_id; run 358 drivers/s390/scsi/zfcp_dbf.c rec->u.run.rec_status = erp->status; run 359 drivers/s390/scsi/zfcp_dbf.c rec->u.run.rec_step = erp->step; run 360 drivers/s390/scsi/zfcp_dbf.c rec->u.run.rec_action = erp->type; run 363 drivers/s390/scsi/zfcp_dbf.c rec->u.run.rec_count = run 366 drivers/s390/scsi/zfcp_dbf.c rec->u.run.rec_count = atomic_read(&erp->port->erp_counter); run 368 drivers/s390/scsi/zfcp_dbf.c rec->u.run.rec_count = atomic_read(&erp->adapter->erp_counter); run 410 drivers/s390/scsi/zfcp_dbf.c rec->u.run.fsf_req_id = req_id; run 411 drivers/s390/scsi/zfcp_dbf.c rec->u.run.rec_status = ~0; run 412 drivers/s390/scsi/zfcp_dbf.c rec->u.run.rec_step = ~0; run 413 drivers/s390/scsi/zfcp_dbf.c rec->u.run.rec_action = ~0; run 414 drivers/s390/scsi/zfcp_dbf.c rec->u.run.rec_count = ~0; run 91 drivers/s390/scsi/zfcp_dbf.h struct zfcp_dbf_rec_running run; run 598 drivers/scsi/aha152x.c void (*run)(struct Scsi_Host *); run 1346 drivers/scsi/aha152x.c INIT_WORK(&aha152x_tq, run); run 2361 drivers/scsi/aha152x.c if(states[STATE].run) run 2362 drivers/scsi/aha152x.c states[STATE].run(shpnt); run 1222 drivers/staging/comedi/drivers/amplc_pci230.c bool run; run 1225 drivers/staging/comedi/drivers/amplc_pci230.c run = pci230_handle_ao_fifo(dev, s); run 1227 drivers/staging/comedi/drivers/amplc_pci230.c if (!run) { run 334 drivers/staging/comedi/drivers/s626.h * Start ADC: run A1 run 194 drivers/staging/media/hantro/hantro_drv.c ctx->codec_ops->run(ctx); run 128 drivers/staging/media/hantro/hantro_hw.h void (*run)(struct hantro_ctx *ctx); run 173 drivers/staging/media/hantro/rk3288_vpu_hw.c .run = hantro_h1_jpeg_enc_run, run 179 drivers/staging/media/hantro/rk3288_vpu_hw.c .run = hantro_g1_h264_dec_run, run 185 drivers/staging/media/hantro/rk3288_vpu_hw.c .run = hantro_g1_mpeg2_dec_run, run 191 drivers/staging/media/hantro/rk3288_vpu_hw.c .run = hantro_g1_vp8_dec_run, run 159 drivers/staging/media/hantro/rk3399_vpu_hw.c .run = rk3399_vpu_jpeg_enc_run, run 165 drivers/staging/media/hantro/rk3399_vpu_hw.c .run = rk3399_vpu_mpeg2_dec_run, run 171 drivers/staging/media/hantro/rk3399_vpu_hw.c .run = rk3399_vpu_vp8_dec_run, run 88 drivers/staging/media/imx/imx-media-csc-scaler.c static void ipu_ic_pp_complete(struct ipu_image_convert_run *run, void *_ctx) run 102 drivers/staging/media/imx/imx-media-csc-scaler.c v4l2_m2m_buf_done(src_buf, run->status ? VB2_BUF_STATE_ERROR : run 104 drivers/staging/media/imx/imx-media-csc-scaler.c v4l2_m2m_buf_done(dst_buf, run->status ? VB2_BUF_STATE_ERROR : run 108 drivers/staging/media/imx/imx-media-csc-scaler.c kfree(run); run 116 drivers/staging/media/imx/imx-media-csc-scaler.c struct ipu_image_convert_run *run; run 122 drivers/staging/media/imx/imx-media-csc-scaler.c run = kzalloc(sizeof(*run), GFP_KERNEL); run 123 drivers/staging/media/imx/imx-media-csc-scaler.c if (!run) run 126 drivers/staging/media/imx/imx-media-csc-scaler.c run->ctx = ctx->icc; run 127 drivers/staging/media/imx/imx-media-csc-scaler.c run->in_phys = vb2_dma_contig_plane_dma_addr(&src_buf->vb2_buf, 0); run 128 drivers/staging/media/imx/imx-media-csc-scaler.c run->out_phys = vb2_dma_contig_plane_dma_addr(&dst_buf->vb2_buf, 0); run 130 drivers/staging/media/imx/imx-media-csc-scaler.c ret = ipu_image_convert_queue(run); run 120 drivers/staging/media/sunxi/cedrus/cedrus.h void (*setup)(struct cedrus_ctx *ctx, struct cedrus_run *run); run 29 drivers/staging/media/sunxi/cedrus/cedrus_dec.c struct cedrus_run run = {}; run 32 drivers/staging/media/sunxi/cedrus/cedrus_dec.c run.src = v4l2_m2m_next_src_buf(ctx->fh.m2m_ctx); run 33 drivers/staging/media/sunxi/cedrus/cedrus_dec.c run.dst = v4l2_m2m_next_dst_buf(ctx->fh.m2m_ctx); run 36 drivers/staging/media/sunxi/cedrus/cedrus_dec.c src_req = run.src->vb2_buf.req_obj.req; run 43 drivers/staging/media/sunxi/cedrus/cedrus_dec.c run.mpeg2.slice_params = cedrus_find_control_data(ctx, run 45 drivers/staging/media/sunxi/cedrus/cedrus_dec.c run.mpeg2.quantization = cedrus_find_control_data(ctx, run 50 drivers/staging/media/sunxi/cedrus/cedrus_dec.c run.h264.decode_params = cedrus_find_control_data(ctx, run 52 drivers/staging/media/sunxi/cedrus/cedrus_dec.c run.h264.pps = cedrus_find_control_data(ctx, run 54 drivers/staging/media/sunxi/cedrus/cedrus_dec.c run.h264.scaling_matrix = cedrus_find_control_data(ctx, run 56 drivers/staging/media/sunxi/cedrus/cedrus_dec.c run.h264.slice_params = cedrus_find_control_data(ctx, run 58 drivers/staging/media/sunxi/cedrus/cedrus_dec.c run.h264.sps = cedrus_find_control_data(ctx, run 66 drivers/staging/media/sunxi/cedrus/cedrus_dec.c v4l2_m2m_buf_copy_metadata(run.src, run.dst, true); run 68 drivers/staging/media/sunxi/cedrus/cedrus_dec.c dev->dec_ops[ctx->current_codec]->setup(ctx, &run); run 93 drivers/staging/media/sunxi/cedrus/cedrus_h264.c struct cedrus_run *run) run 96 drivers/staging/media/sunxi/cedrus/cedrus_h264.c const struct v4l2_ctrl_h264_decode_params *decode = run->h264.decode_params; run 97 drivers/staging/media/sunxi/cedrus/cedrus_h264.c const struct v4l2_ctrl_h264_slice_params *slice = run->h264.slice_params; run 98 drivers/staging/media/sunxi/cedrus/cedrus_h264.c const struct v4l2_ctrl_h264_sps *sps = run->h264.sps; run 143 drivers/staging/media/sunxi/cedrus/cedrus_h264.c output_buf = vb2_to_cedrus_buffer(&run->dst->vb2_buf); run 167 drivers/staging/media/sunxi/cedrus/cedrus_h264.c struct cedrus_run *run, run 171 drivers/staging/media/sunxi/cedrus/cedrus_h264.c const struct v4l2_ctrl_h264_decode_params *decode = run->h264.decode_params; run 214 drivers/staging/media/sunxi/cedrus/cedrus_h264.c struct cedrus_run *run) run 216 drivers/staging/media/sunxi/cedrus/cedrus_h264.c const struct v4l2_ctrl_h264_slice_params *slice = run->h264.slice_params; run 218 drivers/staging/media/sunxi/cedrus/cedrus_h264.c _cedrus_write_ref_list(ctx, run, run 225 drivers/staging/media/sunxi/cedrus/cedrus_h264.c struct cedrus_run *run) run 227 drivers/staging/media/sunxi/cedrus/cedrus_h264.c const struct v4l2_ctrl_h264_slice_params *slice = run->h264.slice_params; run 229 drivers/staging/media/sunxi/cedrus/cedrus_h264.c _cedrus_write_ref_list(ctx, run, run 236 drivers/staging/media/sunxi/cedrus/cedrus_h264.c struct cedrus_run *run) run 239 drivers/staging/media/sunxi/cedrus/cedrus_h264.c run->h264.scaling_matrix; run 256 drivers/staging/media/sunxi/cedrus/cedrus_h264.c struct cedrus_run *run) run 259 drivers/staging/media/sunxi/cedrus/cedrus_h264.c run->h264.slice_params; run 297 drivers/staging/media/sunxi/cedrus/cedrus_h264.c struct cedrus_run *run) run 299 drivers/staging/media/sunxi/cedrus/cedrus_h264.c const struct v4l2_ctrl_h264_decode_params *decode = run->h264.decode_params; run 300 drivers/staging/media/sunxi/cedrus/cedrus_h264.c const struct v4l2_ctrl_h264_slice_params *slice = run->h264.slice_params; run 301 drivers/staging/media/sunxi/cedrus/cedrus_h264.c const struct v4l2_ctrl_h264_pps *pps = run->h264.pps; run 302 drivers/staging/media/sunxi/cedrus/cedrus_h264.c const struct v4l2_ctrl_h264_sps *sps = run->h264.sps; run 303 drivers/staging/media/sunxi/cedrus/cedrus_h264.c struct vb2_buffer *src_buf = &run->src->vb2_buf; run 335 drivers/staging/media/sunxi/cedrus/cedrus_h264.c cedrus_write_pred_weight_table(ctx, run); run 340 drivers/staging/media/sunxi/cedrus/cedrus_h264.c cedrus_write_ref_list0(ctx, run); run 343 drivers/staging/media/sunxi/cedrus/cedrus_h264.c cedrus_write_ref_list1(ctx, run); run 450 drivers/staging/media/sunxi/cedrus/cedrus_h264.c struct cedrus_run *run) run 462 drivers/staging/media/sunxi/cedrus/cedrus_h264.c cedrus_write_scaling_lists(ctx, run); run 463 drivers/staging/media/sunxi/cedrus/cedrus_h264.c cedrus_write_frame_list(ctx, run); run 465 drivers/staging/media/sunxi/cedrus/cedrus_h264.c cedrus_set_params(ctx, run); run 75 drivers/staging/media/sunxi/cedrus/cedrus_mpeg2.c static void cedrus_mpeg2_setup(struct cedrus_ctx *ctx, struct cedrus_run *run) run 92 drivers/staging/media/sunxi/cedrus/cedrus_mpeg2.c slice_params = run->mpeg2.slice_params; run 96 drivers/staging/media/sunxi/cedrus/cedrus_mpeg2.c quantization = run->mpeg2.quantization; run 181 drivers/staging/media/sunxi/cedrus/cedrus_mpeg2.c dst_luma_addr = cedrus_dst_buf_addr(ctx, run->dst->vb2_buf.index, 0); run 182 drivers/staging/media/sunxi/cedrus/cedrus_mpeg2.c dst_chroma_addr = cedrus_dst_buf_addr(ctx, run->dst->vb2_buf.index, 1); run 197 drivers/staging/media/sunxi/cedrus/cedrus_mpeg2.c src_buf_addr = vb2_dma_contig_plane_dma_addr(&run->src->vb2_buf, 0); run 234 drivers/tty/tty_audit.c size_t run; run 236 drivers/tty/tty_audit.c run = N_TTY_BUF_SIZE - buf->valid; run 237 drivers/tty/tty_audit.c if (run > size) run 238 drivers/tty/tty_audit.c run = size; run 239 drivers/tty/tty_audit.c memcpy(buf->data + buf->valid, data, run); run 240 drivers/tty/tty_audit.c buf->valid += run; run 241 drivers/tty/tty_audit.c data += run; run 242 drivers/tty/tty_audit.c size -= run; run 409 drivers/usb/phy/phy-mv-usb.c run: run 475 drivers/usb/phy/phy-mv-usb.c goto run; run 26 fs/befs/datastream.c befs_blocknr_t blockno, befs_block_run *run); run 31 fs/befs/datastream.c befs_block_run *run); run 36 fs/befs/datastream.c befs_block_run *run); run 53 fs/befs/datastream.c befs_block_run run; run 61 fs/befs/datastream.c if (befs_fblock2brun(sb, ds, block, &run) != BEFS_OK) { run 67 fs/befs/datastream.c bh = befs_bread_iaddr(sb, run); run 96 fs/befs/datastream.c befs_blocknr_t fblock, befs_block_run *run) run 102 fs/befs/datastream.c err = befs_find_brun_direct(sb, data, fblock, run); run 105 fs/befs/datastream.c err = befs_find_brun_indirect(sb, data, fblock, run); run 108 fs/befs/datastream.c err = befs_find_brun_dblindirect(sb, data, fblock, run); run 252 fs/befs/datastream.c befs_blocknr_t blockno, befs_block_run *run) run 265 fs/befs/datastream.c run->allocation_group = array[i].allocation_group; run 266 fs/befs/datastream.c run->start = array[i].start + offset; run 267 fs/befs/datastream.c run->len = array[i].len - offset; run 308 fs/befs/datastream.c befs_block_run *run) run 344 fs/befs/datastream.c run->allocation_group = run 346 fs/befs/datastream.c run->start = run 348 fs/befs/datastream.c run->len = run 417 fs/befs/datastream.c befs_block_run *run) run 509 fs/befs/datastream.c *run = fsrun_to_cpu(sb, iaddr_array[block_indx]); run 517 fs/befs/datastream.c run->start += offset; run 518 fs/befs/datastream.c run->len -= offset; run 12 fs/befs/datastream.h befs_blocknr_t fblock, befs_block_run *run); run 213 fs/befs/debug.c befs_dump_run(const struct super_block *sb, befs_disk_block_run run) run 217 fs/befs/debug.c befs_block_run n = fsrun_to_cpu(sb, run); run 74 fs/befs/endian.h befs_block_run run; run 77 fs/befs/endian.h run.allocation_group = le32_to_cpu((__force __le32)n.allocation_group); run 78 fs/befs/endian.h run.start = le16_to_cpu((__force __le16)n.start); run 79 fs/befs/endian.h run.len = le16_to_cpu((__force __le16)n.len); run 81 fs/befs/endian.h run.allocation_group = be32_to_cpu((__force __be32)n.allocation_group); run 82 fs/befs/endian.h run.start = be16_to_cpu((__force __be16)n.start); run 83 fs/befs/endian.h run.len = be16_to_cpu((__force __be16)n.len); run 85 fs/befs/endian.h return run; run 91 fs/befs/endian.h befs_disk_block_run run; run 94 fs/befs/endian.h run.allocation_group = cpu_to_le32(n.allocation_group); run 95 fs/befs/endian.h run.start = cpu_to_le16(n.start); run 96 fs/befs/endian.h run.len = cpu_to_le16(n.len); run 98 fs/befs/endian.h run.allocation_group = cpu_to_be32(n.allocation_group); run 99 fs/befs/endian.h run.start = cpu_to_be16(n.start); run 100 fs/befs/endian.h run.len = cpu_to_be16(n.len); run 102 fs/befs/endian.h return run; run 137 fs/befs/linuxvfs.c befs_block_run run = BAD_IADDR; run 150 fs/befs/linuxvfs.c res = befs_fblock2brun(sb, ds, block, &run); run 159 fs/befs/linuxvfs.c disk_off = (ulong) iaddr2blockno(sb, &run); run 1275 fs/gfs2/dir.c int run = 0; run 1297 fs/gfs2/dir.c if (*copied && !run) run 1299 fs/gfs2/dir.c run = 1; run 1301 fs/gfs2/dir.c run = 0; run 427 fs/jbd2/commit.c stats.run.rs_wait = commit_transaction->t_max_wait; run 428 fs/jbd2/commit.c stats.run.rs_request_delay = 0; run 429 fs/jbd2/commit.c stats.run.rs_locked = jiffies; run 431 fs/jbd2/commit.c stats.run.rs_request_delay = run 433 fs/jbd2/commit.c stats.run.rs_locked); run 434 fs/jbd2/commit.c stats.run.rs_running = jbd2_time_diff(commit_transaction->t_start, run 435 fs/jbd2/commit.c stats.run.rs_locked); run 523 fs/jbd2/commit.c stats.run.rs_flushing = jiffies; run 524 fs/jbd2/commit.c stats.run.rs_locked = jbd2_time_diff(stats.run.rs_locked, run 525 fs/jbd2/commit.c stats.run.rs_flushing); run 560 fs/jbd2/commit.c stats.run.rs_logging = jiffies; run 561 fs/jbd2/commit.c stats.run.rs_flushing = jbd2_time_diff(stats.run.rs_flushing, run 562 fs/jbd2/commit.c stats.run.rs_logging); run 563 fs/jbd2/commit.c stats.run.rs_blocks = run 565 fs/jbd2/commit.c stats.run.rs_blocks_logged = 0; run 816 fs/jbd2/commit.c stats.run.rs_blocks_logged++; run 862 fs/jbd2/commit.c stats.run.rs_blocks_logged++; run 884 fs/jbd2/commit.c stats.run.rs_blocks_logged++; run 1085 fs/jbd2/commit.c stats.run.rs_logging = jbd2_time_diff(stats.run.rs_logging, run 1092 fs/jbd2/commit.c stats.run.rs_handle_count = run 1095 fs/jbd2/commit.c commit_transaction->t_tid, &stats.run); run 1142 fs/jbd2/commit.c journal->j_stats.run.rs_wait += stats.run.rs_wait; run 1143 fs/jbd2/commit.c journal->j_stats.run.rs_request_delay += stats.run.rs_request_delay; run 1144 fs/jbd2/commit.c journal->j_stats.run.rs_running += stats.run.rs_running; run 1145 fs/jbd2/commit.c journal->j_stats.run.rs_locked += stats.run.rs_locked; run 1146 fs/jbd2/commit.c journal->j_stats.run.rs_flushing += stats.run.rs_flushing; run 1147 fs/jbd2/commit.c journal->j_stats.run.rs_logging += stats.run.rs_logging; run 1148 fs/jbd2/commit.c journal->j_stats.run.rs_handle_count += stats.run.rs_handle_count; run 1149 fs/jbd2/commit.c journal->j_stats.run.rs_blocks += stats.run.rs_blocks; run 1150 fs/jbd2/commit.c journal->j_stats.run.rs_blocks_logged += stats.run.rs_blocks_logged; run 1001 fs/jbd2/journal.c jiffies_to_msecs(s->stats->run.rs_wait / s->stats->ts_tid)); run 1004 fs/jbd2/journal.c jiffies_to_msecs(s->stats->run.rs_request_delay / run 1007 fs/jbd2/journal.c jiffies_to_msecs(s->stats->run.rs_running / s->stats->ts_tid)); run 1009 fs/jbd2/journal.c jiffies_to_msecs(s->stats->run.rs_locked / s->stats->ts_tid)); run 1011 fs/jbd2/journal.c jiffies_to_msecs(s->stats->run.rs_flushing / s->stats->ts_tid)); run 1013 fs/jbd2/journal.c jiffies_to_msecs(s->stats->run.rs_logging / s->stats->ts_tid)); run 1017 fs/jbd2/journal.c s->stats->run.rs_handle_count / s->stats->ts_tid); run 1019 fs/jbd2/journal.c s->stats->run.rs_blocks / s->stats->ts_tid); run 1021 fs/jbd2/journal.c s->stats->run.rs_blocks_logged / s->stats->ts_tid); run 139 fs/omfs/bitmap.c int i, run, bit; run 151 fs/omfs/bitmap.c run = count_run(&sbi->s_imap[i], bits_per_entry, run 154 fs/omfs/bitmap.c if (run >= min_request) run 156 fs/omfs/bitmap.c bit += run; run 164 fs/omfs/bitmap.c *return_size = run; run 165 fs/omfs/bitmap.c ret = set_run(sb, i, bits_per_entry, bit, run, 1); run 52 include/linux/blk-mq.h unsigned long run; run 743 include/linux/jbd2.h struct transaction_run_stats_s run; run 279 include/linux/kvm_host.h struct kvm_run *run; run 651 include/linux/phy.h int (*run)(struct phy_device *phydev); run 1172 include/linux/phy.h int (*run)(struct phy_device *)); run 1174 include/linux/phy.h int (*run)(struct phy_device *)); run 1176 include/linux/phy.h int (*run)(struct phy_device *)); run 40 include/video/imx-ipu-image-convert.h typedef void (*ipu_image_convert_cb_t)(struct ipu_image_convert_run *run, run 139 include/video/imx-ipu-image-convert.h int ipu_image_convert_queue(struct ipu_image_convert_run *run); run 489 kernel/stop_machine.c int run; run 492 kernel/stop_machine.c run = !list_empty(&stopper->works); run 494 kernel/stop_machine.c return run; run 2946 mm/ksm.c KSM_ATTR(run); run 900 sound/soc/qcom/qdsp6/q6asm.c struct asm_session_cmd_run_v2 *run; run 905 sound/soc/qcom/qdsp6/q6asm.c pkt_size = APR_HDR_SIZE + sizeof(*run); run 911 sound/soc/qcom/qdsp6/q6asm.c run = p + APR_HDR_SIZE; run 916 sound/soc/qcom/qdsp6/q6asm.c run->flags = flags; run 917 sound/soc/qcom/qdsp6/q6asm.c run->time_lsw = lsw_ts; run 918 sound/soc/qcom/qdsp6/q6asm.c run->time_msw = msw_ts; run 491 sound/soc/sof/core.c if (!sof_ops(sdev) || !sof_ops(sdev)->probe || !sof_ops(sdev)->run || run 370 sound/soc/sof/imx/imx8.c .run = imx8_run, run 74 sound/soc/sof/intel/apl.c .run = hda_dsp_cl_boot_firmware, run 535 sound/soc/sof/intel/bdw.c .run = bdw_run, run 471 sound/soc/sof/intel/byt.c .run = byt_run, run 632 sound/soc/sof/intel/byt.c .run = byt_run, run 691 sound/soc/sof/intel/byt.c .run = byt_run, run 240 sound/soc/sof/intel/cnl.c .run = hda_dsp_cl_boot_firmware, run 253 sound/soc/sof/intel/hda-stream.c u32 run; run 273 sound/soc/sof/intel/hda-stream.c sd_offset, run, run 274 sound/soc/sof/intel/hda-stream.c ((run & dma_start) == dma_start), run 292 sound/soc/sof/intel/hda-stream.c sd_offset, run, run 293 sound/soc/sof/intel/hda-stream.c !(run & dma_start), run 331 sound/soc/sof/intel/hda-stream.c u32 run; run 354 sound/soc/sof/intel/hda-stream.c sd_offset, run, run 355 sound/soc/sof/intel/hda-stream.c !(run & dma_start), run 416 sound/soc/sof/intel/hda-stream.c sd_offset, run, run 417 sound/soc/sof/intel/hda-stream.c !(run & dma_start), run 48 sound/soc/sof/ops.h return sof_ops(sdev)->run(sdev); run 85 sound/soc/sof/sof-priv.h int (*run)(struct snd_sof_dev *sof_dev); /* mandatory */ run 36 tools/firewire/nosy-dump.c static int run = 1; run 127 tools/firewire/nosy-dump.c if (run == 1) { run 128 tools/firewire/nosy-dump.c run = 0; run 973 tools/firewire/nosy-dump.c while (run) { run 1946 tools/perf/builtin-script.c counts->run, run 257 tools/perf/builtin-stat.c count->ena = count->run = val; run 298 tools/perf/builtin-stat.c perf_counts(counter->counts, cpu, thread)->run = 0; run 316 tools/perf/builtin-stat.c count->val, count->ena, count->run); run 296 tools/perf/lib/include/perf/event.h __u64 run; run 18 tools/perf/lib/include/perf/evsel.h uint64_t run; run 202 tools/perf/tests/event-times.c err = !(count.ena == count.run); run 206 tools/perf/tests/event-times.c count.ena, count.run); run 76 tools/perf/tests/stat.c TEST_ASSERT_VAL("wrong ena", st->run == 300); run 86 tools/perf/tests/stat.c count.run = 300; run 161 tools/perf/util/c++/clang.cpp PM.run(*Module); run 1284 tools/perf/util/evsel.c count->run = count->run - tmp.run; run 1293 tools/perf/util/evsel.c if (count->run == 0) { run 1296 tools/perf/util/evsel.c } else if (count->run < count->ena) { run 1298 tools/perf/util/evsel.c count->val = (u64)((double) count->val * count->ena / count->run); run 1316 tools/perf/util/evsel.c u64 val, u64 ena, u64 run) run 1324 tools/perf/util/evsel.c count->run = run; run 1335 tools/perf/util/evsel.c u64 nr, ena = 0, run = 0, i; run 1346 tools/perf/util/evsel.c run = *data++; run 1351 tools/perf/util/evsel.c v[0].value, ena, run); run 1361 tools/perf/util/evsel.c v[i].value, ena, run); run 1382 tools/perf/util/scripting-engines/trace-event-python.c tuple_set_u64(t, n++, count->run); run 899 tools/perf/util/session.c event->stat.run = bswap_64(event->stat.run); run 24 tools/perf/util/stat-display.c u64 run, u64 ena) run 29 tools/perf/util/stat-display.c run, run 31 tools/perf/util/stat-display.c ena ? 100.0 * run / ena : 100.0); run 32 tools/perf/util/stat-display.c } else if (run != ena) { run 33 tools/perf/util/stat-display.c fprintf(config->output, " (%.2f%%)", 100.0 * run / ena); run 393 tools/perf/util/stat-display.c char *prefix, u64 run, u64 ena, double noise, run 433 tools/perf/util/stat-display.c if (run == 0 || ena == 0 || counter->counts->scaled == -1) { run 464 tools/perf/util/stat-display.c print_running(config, run, ena); run 480 tools/perf/util/stat-display.c print_running(config, run, ena); run 488 tools/perf/util/stat-display.c print_running(config, run, ena); run 582 tools/perf/util/stat-display.c u64 ena, run, val; run 607 tools/perf/util/stat-display.c if (counts->ena == 0 || counts->run == 0 || run 610 tools/perf/util/stat-display.c ad->run = 0; run 615 tools/perf/util/stat-display.c ad->run += counts->run; run 626 tools/perf/util/stat-display.c u64 ena, run, val; run 631 tools/perf/util/stat-display.c ad.val = ad.ena = ad.run = 0; run 638 tools/perf/util/stat-display.c run = ad.run; run 649 tools/perf/util/stat-display.c run, ena, 1.0, &rt_stat); run 709 tools/perf/util/stat-display.c u64 ena = 0, run = 0, val = 0; run 714 tools/perf/util/stat-display.c run += perf_counts(counter->counts, cpu, thread)->run; run 730 tools/perf/util/stat-display.c buf[i].run = run; run 766 tools/perf/util/stat-display.c prefix, buf[thread].run, buf[thread].ena, 1.0, run 770 tools/perf/util/stat-display.c prefix, buf[thread].run, buf[thread].ena, 1.0, run 827 tools/perf/util/stat-display.c ad->run += perf_counts(counter->counts, ad->cpu, 0)->run; run 838 tools/perf/util/stat-display.c u64 ena, run, val; run 849 tools/perf/util/stat-display.c run = ad.run; run 855 tools/perf/util/stat-display.c printout(config, cpu, 0, counter, uval, prefix, run, ena, 1.0, run 869 tools/perf/util/stat-display.c u64 ena, run, val; run 885 tools/perf/util/stat-display.c run = perf_counts(counter->counts, cpu, 0)->run; run 888 tools/perf/util/stat-display.c printout(config, cpu, 0, counter, uval, prefix, run, ena, 1.0, run 1062 tools/perf/util/stat-display.c double run = (double) config->walltime_run[idx] / NSEC_PER_SEC; run 1063 tools/perf/util/stat-display.c int h, n = 1 + abs((int) (100.0 * (run - avg)/run) / 5); run 1066 tools/perf/util/stat-display.c precision, run, precision, run - avg); run 170 tools/perf/util/stat.c evsel->prev_raw_counts->aggr.run = 0; run 269 tools/perf/util/stat.c if (!(vals->run && vals->ena)) run 323 tools/perf/util/stat.c aggr->run += count->run; run 361 tools/perf/util/stat.c aggr->val = aggr->ena = aggr->run = 0; run 412 tools/perf/util/stat.c count.run = st->run; run 433 tools/perf/util/stat.c st->val, st->ena, st->run); run 153 tools/perf/util/stat.h u64 run; run 1105 tools/perf/util/synthetic-events.c event.run = count->run; run 799 tools/testing/selftests/bpf/test_sockmap.c goto run; run 827 tools/testing/selftests/bpf/test_sockmap.c run: run 154 tools/testing/selftests/kvm/dirty_log_test.c struct kvm_run *run; run 156 tools/testing/selftests/kvm/dirty_log_test.c run = vcpu_state(vm, VCPU_ID); run 172 tools/testing/selftests/kvm/dirty_log_test.c exit_reason_str(run->exit_reason)); run 94 tools/testing/selftests/kvm/lib/aarch64/ucall.c struct kvm_run *run = vcpu_state(vm, vcpu_id); run 97 tools/testing/selftests/kvm/lib/aarch64/ucall.c if (run->exit_reason == KVM_EXIT_MMIO && run 98 tools/testing/selftests/kvm/lib/aarch64/ucall.c run->mmio.phys_addr == (uint64_t)ucall_exit_mmio_addr) { run 101 tools/testing/selftests/kvm/lib/aarch64/ucall.c TEST_ASSERT(run->mmio.is_write && run->mmio.len == 8, run 103 tools/testing/selftests/kvm/lib/aarch64/ucall.c memcpy(&gva, run->mmio.data, sizeof(gva)); run 247 tools/testing/selftests/kvm/lib/s390x/processor.c struct kvm_run *run; run 267 tools/testing/selftests/kvm/lib/s390x/processor.c run = vcpu_state(vm, vcpuid); run 268 tools/testing/selftests/kvm/lib/s390x/processor.c run->psw_mask = 0x0400000180000000ULL; /* DAT enabled + 64 bit mode */ run 269 tools/testing/selftests/kvm/lib/s390x/processor.c run->psw_addr = (uintptr_t)guest_code; run 38 tools/testing/selftests/kvm/lib/s390x/ucall.c struct kvm_run *run = vcpu_state(vm, vcpu_id); run 41 tools/testing/selftests/kvm/lib/s390x/ucall.c if (run->exit_reason == KVM_EXIT_S390_SIEIC && run 42 tools/testing/selftests/kvm/lib/s390x/ucall.c run->s390_sieic.icptcode == 4 && run 43 tools/testing/selftests/kvm/lib/s390x/ucall.c (run->s390_sieic.ipa >> 8) == 0x83 && /* 0x83 means DIAGNOSE */ run 44 tools/testing/selftests/kvm/lib/s390x/ucall.c (run->s390_sieic.ipb >> 16) == 0x501) { run 45 tools/testing/selftests/kvm/lib/s390x/ucall.c int reg = run->s390_sieic.ipa & 0xf; run 47 tools/testing/selftests/kvm/lib/s390x/ucall.c memcpy(&ucall, addr_gva2hva(vm, run->s.regs.gprs[reg]), run 40 tools/testing/selftests/kvm/lib/x86_64/ucall.c struct kvm_run *run = vcpu_state(vm, vcpu_id); run 43 tools/testing/selftests/kvm/lib/x86_64/ucall.c if (run->exit_reason == KVM_EXIT_IO && run->io.port == UCALL_PIO_PORT) { run 35 tools/testing/selftests/kvm/s390x/memop.c struct kvm_run *run; run 51 tools/testing/selftests/kvm/s390x/memop.c run = vcpu_state(vm, VCPU_ID); run 67 tools/testing/selftests/kvm/s390x/memop.c TEST_ASSERT(run->exit_reason == KVM_EXIT_S390_SIEIC, run 69 tools/testing/selftests/kvm/s390x/memop.c run->exit_reason, run 70 tools/testing/selftests/kvm/s390x/memop.c exit_reason_str(run->exit_reason)); run 149 tools/testing/selftests/kvm/s390x/memop.c run->psw_mask &= ~(3UL << (63 - 17)); run 150 tools/testing/selftests/kvm/s390x/memop.c run->psw_mask |= 1UL << (63 - 17); /* Enable AR mode */ run 160 tools/testing/selftests/kvm/s390x/memop.c run->psw_mask &= ~(3UL << (63 - 17)); /* Disable AR mode */ run 69 tools/testing/selftests/kvm/s390x/sync_regs_test.c struct kvm_run *run; run 86 tools/testing/selftests/kvm/s390x/sync_regs_test.c run = vcpu_state(vm, VCPU_ID); run 89 tools/testing/selftests/kvm/s390x/sync_regs_test.c run->kvm_valid_regs = INVALID_SYNC_FIELD; run 96 tools/testing/selftests/kvm/s390x/sync_regs_test.c run->kvm_valid_regs = INVALID_SYNC_FIELD | TEST_SYNC_FIELDS; run 104 tools/testing/selftests/kvm/s390x/sync_regs_test.c run->kvm_dirty_regs = INVALID_SYNC_FIELD; run 111 tools/testing/selftests/kvm/s390x/sync_regs_test.c run->kvm_dirty_regs = INVALID_SYNC_FIELD | TEST_SYNC_FIELDS; run 119 tools/testing/selftests/kvm/s390x/sync_regs_test.c run->kvm_valid_regs = TEST_SYNC_FIELDS; run 122 tools/testing/selftests/kvm/s390x/sync_regs_test.c TEST_ASSERT(run->exit_reason == KVM_EXIT_S390_SIEIC, run 124 tools/testing/selftests/kvm/s390x/sync_regs_test.c run->exit_reason, run 125 tools/testing/selftests/kvm/s390x/sync_regs_test.c exit_reason_str(run->exit_reason)); run 126 tools/testing/selftests/kvm/s390x/sync_regs_test.c TEST_ASSERT(run->s390_sieic.icptcode == 4 && run 127 tools/testing/selftests/kvm/s390x/sync_regs_test.c (run->s390_sieic.ipa >> 8) == 0x83 && run 128 tools/testing/selftests/kvm/s390x/sync_regs_test.c (run->s390_sieic.ipb >> 16) == 0x501, run 130 tools/testing/selftests/kvm/s390x/sync_regs_test.c run->s390_sieic.icptcode, run->s390_sieic.ipa, run 131 tools/testing/selftests/kvm/s390x/sync_regs_test.c run->s390_sieic.ipb); run 134 tools/testing/selftests/kvm/s390x/sync_regs_test.c compare_regs(®s, &run->s.regs); run 137 tools/testing/selftests/kvm/s390x/sync_regs_test.c compare_sregs(&sregs, &run->s.regs); run 140 tools/testing/selftests/kvm/s390x/sync_regs_test.c run->s.regs.gprs[11] = 0xBAD1DEA; run 141 tools/testing/selftests/kvm/s390x/sync_regs_test.c run->s.regs.acrs[0] = 1 << 11; run 143 tools/testing/selftests/kvm/s390x/sync_regs_test.c run->kvm_valid_regs = TEST_SYNC_FIELDS; run 144 tools/testing/selftests/kvm/s390x/sync_regs_test.c run->kvm_dirty_regs = KVM_SYNC_GPRS | KVM_SYNC_ACRS; run 147 tools/testing/selftests/kvm/s390x/sync_regs_test.c TEST_ASSERT(run->exit_reason == KVM_EXIT_S390_SIEIC, run 149 tools/testing/selftests/kvm/s390x/sync_regs_test.c run->exit_reason, run 150 tools/testing/selftests/kvm/s390x/sync_regs_test.c exit_reason_str(run->exit_reason)); run 151 tools/testing/selftests/kvm/s390x/sync_regs_test.c TEST_ASSERT(run->s.regs.gprs[11] == 0xBAD1DEA + 1, run 153 tools/testing/selftests/kvm/s390x/sync_regs_test.c run->s.regs.gprs[11]); run 154 tools/testing/selftests/kvm/s390x/sync_regs_test.c TEST_ASSERT(run->s.regs.acrs[0] == 1 << 11, run 156 tools/testing/selftests/kvm/s390x/sync_regs_test.c run->s.regs.acrs[0]); run 159 tools/testing/selftests/kvm/s390x/sync_regs_test.c compare_regs(®s, &run->s.regs); run 162 tools/testing/selftests/kvm/s390x/sync_regs_test.c compare_sregs(&sregs, &run->s.regs); run 167 tools/testing/selftests/kvm/s390x/sync_regs_test.c run->kvm_valid_regs = TEST_SYNC_FIELDS; run 168 tools/testing/selftests/kvm/s390x/sync_regs_test.c run->kvm_dirty_regs = 0; run 169 tools/testing/selftests/kvm/s390x/sync_regs_test.c run->s.regs.gprs[11] = 0xDEADBEEF; run 172 tools/testing/selftests/kvm/s390x/sync_regs_test.c TEST_ASSERT(run->exit_reason == KVM_EXIT_S390_SIEIC, run 174 tools/testing/selftests/kvm/s390x/sync_regs_test.c run->exit_reason, run 175 tools/testing/selftests/kvm/s390x/sync_regs_test.c exit_reason_str(run->exit_reason)); run 176 tools/testing/selftests/kvm/s390x/sync_regs_test.c TEST_ASSERT(run->s.regs.gprs[11] != 0xDEADBEEF, run 178 tools/testing/selftests/kvm/s390x/sync_regs_test.c run->s.regs.gprs[11]); run 66 tools/testing/selftests/kvm/x86_64/cr4_cpuid_sync_test.c struct kvm_run *run; run 85 tools/testing/selftests/kvm/x86_64/cr4_cpuid_sync_test.c run = vcpu_state(vm, VCPU_ID); run 91 tools/testing/selftests/kvm/x86_64/cr4_cpuid_sync_test.c TEST_ASSERT(run->exit_reason == KVM_EXIT_IO, run 93 tools/testing/selftests/kvm/x86_64/cr4_cpuid_sync_test.c run->exit_reason, run 94 tools/testing/selftests/kvm/x86_64/cr4_cpuid_sync_test.c exit_reason_str(run->exit_reason)); run 78 tools/testing/selftests/kvm/x86_64/evmcs_test.c struct kvm_run *run; run 96 tools/testing/selftests/kvm/x86_64/evmcs_test.c run = vcpu_state(vm, VCPU_ID); run 105 tools/testing/selftests/kvm/x86_64/evmcs_test.c TEST_ASSERT(run->exit_reason == KVM_EXIT_IO, run 107 tools/testing/selftests/kvm/x86_64/evmcs_test.c stage, run->exit_reason, run 108 tools/testing/selftests/kvm/x86_64/evmcs_test.c exit_reason_str(run->exit_reason)); run 140 tools/testing/selftests/kvm/x86_64/evmcs_test.c run = vcpu_state(vm, VCPU_ID); run 36 tools/testing/selftests/kvm/x86_64/mmio_warning_test.c struct kvm_run *run; run 44 tools/testing/selftests/kvm/x86_64/mmio_warning_test.c struct kvm_run *run = tc->run; run 48 tools/testing/selftests/kvm/x86_64/mmio_warning_test.c res, run->exit_reason, run->internal.suberror); run 57 tools/testing/selftests/kvm/x86_64/mmio_warning_test.c struct kvm_run *run; run 66 tools/testing/selftests/kvm/x86_64/mmio_warning_test.c run = (struct kvm_run *)mmap(0, 4096, PROT_READ|PROT_WRITE, MAP_SHARED, run 69 tools/testing/selftests/kvm/x86_64/mmio_warning_test.c tc.run = run; run 50 tools/testing/selftests/kvm/x86_64/platform_info_test.c struct kvm_run *run = vcpu_state(vm, VCPU_ID); run 55 tools/testing/selftests/kvm/x86_64/platform_info_test.c TEST_ASSERT(run->exit_reason == KVM_EXIT_IO, run 57 tools/testing/selftests/kvm/x86_64/platform_info_test.c run->exit_reason, run 58 tools/testing/selftests/kvm/x86_64/platform_info_test.c exit_reason_str(run->exit_reason)); run 71 tools/testing/selftests/kvm/x86_64/platform_info_test.c struct kvm_run *run = vcpu_state(vm, VCPU_ID); run 75 tools/testing/selftests/kvm/x86_64/platform_info_test.c TEST_ASSERT(run->exit_reason == KVM_EXIT_SHUTDOWN, run 77 tools/testing/selftests/kvm/x86_64/platform_info_test.c run->exit_reason, run 78 tools/testing/selftests/kvm/x86_64/platform_info_test.c exit_reason_str(run->exit_reason)); run 94 tools/testing/selftests/kvm/x86_64/smm_test.c struct kvm_run *run; run 103 tools/testing/selftests/kvm/x86_64/smm_test.c run = vcpu_state(vm, VCPU_ID); run 126 tools/testing/selftests/kvm/x86_64/smm_test.c TEST_ASSERT(run->exit_reason == KVM_EXIT_IO, run 128 tools/testing/selftests/kvm/x86_64/smm_test.c stage, run->exit_reason, run 129 tools/testing/selftests/kvm/x86_64/smm_test.c exit_reason_str(run->exit_reason)); run 150 tools/testing/selftests/kvm/x86_64/smm_test.c run = vcpu_state(vm, VCPU_ID); run 126 tools/testing/selftests/kvm/x86_64/state_test.c struct kvm_run *run; run 134 tools/testing/selftests/kvm/x86_64/state_test.c run = vcpu_state(vm, VCPU_ID); run 148 tools/testing/selftests/kvm/x86_64/state_test.c TEST_ASSERT(run->exit_reason == KVM_EXIT_IO, run 150 tools/testing/selftests/kvm/x86_64/state_test.c stage, run->exit_reason, run 151 tools/testing/selftests/kvm/x86_64/state_test.c exit_reason_str(run->exit_reason)); run 182 tools/testing/selftests/kvm/x86_64/state_test.c run = vcpu_state(vm, VCPU_ID); run 83 tools/testing/selftests/kvm/x86_64/sync_regs_test.c struct kvm_run *run; run 105 tools/testing/selftests/kvm/x86_64/sync_regs_test.c run = vcpu_state(vm, VCPU_ID); run 108 tools/testing/selftests/kvm/x86_64/sync_regs_test.c run->kvm_valid_regs = INVALID_SYNC_FIELD; run 115 tools/testing/selftests/kvm/x86_64/sync_regs_test.c run->kvm_valid_regs = INVALID_SYNC_FIELD | TEST_SYNC_FIELDS; run 123 tools/testing/selftests/kvm/x86_64/sync_regs_test.c run->kvm_dirty_regs = INVALID_SYNC_FIELD; run 130 tools/testing/selftests/kvm/x86_64/sync_regs_test.c run->kvm_dirty_regs = INVALID_SYNC_FIELD | TEST_SYNC_FIELDS; run 139 tools/testing/selftests/kvm/x86_64/sync_regs_test.c run->kvm_valid_regs = TEST_SYNC_FIELDS; run 141 tools/testing/selftests/kvm/x86_64/sync_regs_test.c TEST_ASSERT(run->exit_reason == KVM_EXIT_IO, run 143 tools/testing/selftests/kvm/x86_64/sync_regs_test.c run->exit_reason, run 144 tools/testing/selftests/kvm/x86_64/sync_regs_test.c exit_reason_str(run->exit_reason)); run 147 tools/testing/selftests/kvm/x86_64/sync_regs_test.c compare_regs(®s, &run->s.regs.regs); run 150 tools/testing/selftests/kvm/x86_64/sync_regs_test.c compare_sregs(&sregs, &run->s.regs.sregs); run 153 tools/testing/selftests/kvm/x86_64/sync_regs_test.c compare_vcpu_events(&events, &run->s.regs.events); run 156 tools/testing/selftests/kvm/x86_64/sync_regs_test.c run->s.regs.regs.rbx = 0xBAD1DEA; run 157 tools/testing/selftests/kvm/x86_64/sync_regs_test.c run->s.regs.sregs.apic_base = 1 << 11; run 160 tools/testing/selftests/kvm/x86_64/sync_regs_test.c run->kvm_valid_regs = TEST_SYNC_FIELDS; run 161 tools/testing/selftests/kvm/x86_64/sync_regs_test.c run->kvm_dirty_regs = KVM_SYNC_X86_REGS | KVM_SYNC_X86_SREGS; run 163 tools/testing/selftests/kvm/x86_64/sync_regs_test.c TEST_ASSERT(run->exit_reason == KVM_EXIT_IO, run 165 tools/testing/selftests/kvm/x86_64/sync_regs_test.c run->exit_reason, run 166 tools/testing/selftests/kvm/x86_64/sync_regs_test.c exit_reason_str(run->exit_reason)); run 167 tools/testing/selftests/kvm/x86_64/sync_regs_test.c TEST_ASSERT(run->s.regs.regs.rbx == 0xBAD1DEA + 1, run 169 tools/testing/selftests/kvm/x86_64/sync_regs_test.c run->s.regs.regs.rbx); run 170 tools/testing/selftests/kvm/x86_64/sync_regs_test.c TEST_ASSERT(run->s.regs.sregs.apic_base == 1 << 11, run 172 tools/testing/selftests/kvm/x86_64/sync_regs_test.c run->s.regs.sregs.apic_base); run 175 tools/testing/selftests/kvm/x86_64/sync_regs_test.c compare_regs(®s, &run->s.regs.regs); run 178 tools/testing/selftests/kvm/x86_64/sync_regs_test.c compare_sregs(&sregs, &run->s.regs.sregs); run 181 tools/testing/selftests/kvm/x86_64/sync_regs_test.c compare_vcpu_events(&events, &run->s.regs.events); run 186 tools/testing/selftests/kvm/x86_64/sync_regs_test.c run->kvm_valid_regs = TEST_SYNC_FIELDS; run 187 tools/testing/selftests/kvm/x86_64/sync_regs_test.c run->kvm_dirty_regs = 0; run 188 tools/testing/selftests/kvm/x86_64/sync_regs_test.c run->s.regs.regs.rbx = 0xDEADBEEF; run 190 tools/testing/selftests/kvm/x86_64/sync_regs_test.c TEST_ASSERT(run->exit_reason == KVM_EXIT_IO, run 192 tools/testing/selftests/kvm/x86_64/sync_regs_test.c run->exit_reason, run 193 tools/testing/selftests/kvm/x86_64/sync_regs_test.c exit_reason_str(run->exit_reason)); run 194 tools/testing/selftests/kvm/x86_64/sync_regs_test.c TEST_ASSERT(run->s.regs.regs.rbx != 0xDEADBEEF, run 196 tools/testing/selftests/kvm/x86_64/sync_regs_test.c run->s.regs.regs.rbx); run 202 tools/testing/selftests/kvm/x86_64/sync_regs_test.c run->kvm_valid_regs = 0; run 203 tools/testing/selftests/kvm/x86_64/sync_regs_test.c run->kvm_dirty_regs = 0; run 204 tools/testing/selftests/kvm/x86_64/sync_regs_test.c run->s.regs.regs.rbx = 0xAAAA; run 208 tools/testing/selftests/kvm/x86_64/sync_regs_test.c TEST_ASSERT(run->exit_reason == KVM_EXIT_IO, run 210 tools/testing/selftests/kvm/x86_64/sync_regs_test.c run->exit_reason, run 211 tools/testing/selftests/kvm/x86_64/sync_regs_test.c exit_reason_str(run->exit_reason)); run 212 tools/testing/selftests/kvm/x86_64/sync_regs_test.c TEST_ASSERT(run->s.regs.regs.rbx == 0xAAAA, run 214 tools/testing/selftests/kvm/x86_64/sync_regs_test.c run->s.regs.regs.rbx); run 224 tools/testing/selftests/kvm/x86_64/sync_regs_test.c run->kvm_valid_regs = 0; run 225 tools/testing/selftests/kvm/x86_64/sync_regs_test.c run->kvm_dirty_regs = TEST_SYNC_FIELDS; run 226 tools/testing/selftests/kvm/x86_64/sync_regs_test.c run->s.regs.regs.rbx = 0xBBBB; run 228 tools/testing/selftests/kvm/x86_64/sync_regs_test.c TEST_ASSERT(run->exit_reason == KVM_EXIT_IO, run 230 tools/testing/selftests/kvm/x86_64/sync_regs_test.c run->exit_reason, run 231 tools/testing/selftests/kvm/x86_64/sync_regs_test.c exit_reason_str(run->exit_reason)); run 232 tools/testing/selftests/kvm/x86_64/sync_regs_test.c TEST_ASSERT(run->s.regs.regs.rbx == 0xBBBB, run 234 tools/testing/selftests/kvm/x86_64/sync_regs_test.c run->s.regs.regs.rbx); run 67 tools/testing/selftests/kvm/x86_64/vmx_close_while_nested_test.c volatile struct kvm_run *run = vcpu_state(vm, VCPU_ID); run 71 tools/testing/selftests/kvm/x86_64/vmx_close_while_nested_test.c TEST_ASSERT(run->exit_reason == KVM_EXIT_IO, run 73 tools/testing/selftests/kvm/x86_64/vmx_close_while_nested_test.c run->exit_reason, run 74 tools/testing/selftests/kvm/x86_64/vmx_close_while_nested_test.c exit_reason_str(run->exit_reason)); run 76 tools/testing/selftests/kvm/x86_64/vmx_close_while_nested_test.c if (run->io.port == PORT_L0_EXIT) run 77 tools/testing/selftests/kvm/x86_64/vmx_dirty_log_test.c struct kvm_run *run; run 88 tools/testing/selftests/kvm/x86_64/vmx_dirty_log_test.c run = vcpu_state(vm, VCPU_ID); run 122 tools/testing/selftests/kvm/x86_64/vmx_dirty_log_test.c TEST_ASSERT(run->exit_reason == KVM_EXIT_IO, run 124 tools/testing/selftests/kvm/x86_64/vmx_dirty_log_test.c run->exit_reason, run 125 tools/testing/selftests/kvm/x86_64/vmx_dirty_log_test.c exit_reason_str(run->exit_reason)); run 142 tools/testing/selftests/kvm/x86_64/vmx_tsc_adjust_test.c volatile struct kvm_run *run = vcpu_state(vm, VCPU_ID); run 146 tools/testing/selftests/kvm/x86_64/vmx_tsc_adjust_test.c TEST_ASSERT(run->exit_reason == KVM_EXIT_IO, run 148 tools/testing/selftests/kvm/x86_64/vmx_tsc_adjust_test.c run->exit_reason, run 149 tools/testing/selftests/kvm/x86_64/vmx_tsc_adjust_test.c exit_reason_str(run->exit_reason)); run 659 tools/testing/vsock/vsock_diag_test.c void (*run)(unsigned int peer_cid); run 665 tools/testing/vsock/vsock_diag_test.c run = test_cases[i].run_client; run 667 tools/testing/vsock/vsock_diag_test.c run = test_cases[i].run_server; run 669 tools/testing/vsock/vsock_diag_test.c if (run) run 670 tools/testing/vsock/vsock_diag_test.c run(peer_cid); run 283 virt/kvm/arm/arch_timer.c struct kvm_sync_regs *regs = &vcpu->run->s.regs; run 557 virt/kvm/arm/arch_timer.c struct kvm_sync_regs *sregs = &vcpu->run->s.regs; run 682 virt/kvm/arm/arm.c int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu, struct kvm_run *run) run 693 virt/kvm/arm/arm.c if (run->exit_reason == KVM_EXIT_MMIO) { run 694 virt/kvm/arm/arm.c ret = kvm_handle_mmio_return(vcpu, vcpu->run); run 699 virt/kvm/arm/arm.c if (run->immediate_exit) run 707 virt/kvm/arm/arm.c run->exit_reason = KVM_EXIT_UNKNOWN; run 737 virt/kvm/arm/arm.c run->exit_reason = KVM_EXIT_INTR; run 751 virt/kvm/arm/arm.c run->exit_reason = KVM_EXIT_INTR; run 848 virt/kvm/arm/arm.c handle_exit_early(vcpu, run, ret); run 852 virt/kvm/arm/arm.c ret = handle_exit(vcpu, run, ret); run 83 virt/kvm/arm/mmio.c int kvm_handle_mmio_return(struct kvm_vcpu *vcpu, struct kvm_run *run) run 95 virt/kvm/arm/mmio.c if (!run->mmio.is_write) { run 96 virt/kvm/arm/mmio.c len = run->mmio.len; run 100 virt/kvm/arm/mmio.c data = kvm_mmio_read_buf(run->mmio.data, len); run 111 virt/kvm/arm/mmio.c trace_kvm_mmio(KVM_TRACE_MMIO_READ, len, run->mmio.phys_addr, run 156 virt/kvm/arm/mmio.c int io_mem_abort(struct kvm_vcpu *vcpu, struct kvm_run *run, run 200 virt/kvm/arm/mmio.c run->mmio.is_write = is_write; run 201 virt/kvm/arm/mmio.c run->mmio.phys_addr = fault_ipa; run 202 virt/kvm/arm/mmio.c run->mmio.len = len; run 208 virt/kvm/arm/mmio.c memcpy(run->mmio.data, data_buf, len); run 210 virt/kvm/arm/mmio.c kvm_handle_mmio_return(vcpu, run); run 215 virt/kvm/arm/mmio.c memcpy(run->mmio.data, data_buf, len); run 217 virt/kvm/arm/mmio.c run->exit_reason = KVM_EXIT_MMIO; run 1908 virt/kvm/arm/mmu.c int kvm_handle_guest_abort(struct kvm_vcpu *vcpu, struct kvm_run *run) run 1987 virt/kvm/arm/mmu.c ret = io_mem_abort(vcpu, run, fault_ipa); run 392 virt/kvm/arm/pmu.c struct kvm_sync_regs *sregs = &vcpu->run->s.regs; run 406 virt/kvm/arm/pmu.c struct kvm_sync_regs *regs = &vcpu->run->s.regs; run 206 virt/kvm/arm/psci.c memset(&vcpu->run->system_event, 0, sizeof(vcpu->run->system_event)); run 207 virt/kvm/arm/psci.c vcpu->run->system_event.type = type; run 208 virt/kvm/arm/psci.c vcpu->run->exit_reason = KVM_EXIT_SYSTEM_EVENT; run 345 virt/kvm/kvm_main.c vcpu->run = page_address(page); run 358 virt/kvm/kvm_main.c free_page((unsigned long)vcpu->run); run 373 virt/kvm/kvm_main.c free_page((unsigned long)vcpu->run); run 2717 virt/kvm/kvm_main.c page = virt_to_page(vcpu->run); run 2919 virt/kvm/kvm_main.c r = kvm_arch_vcpu_ioctl_run(vcpu, vcpu->run); run 2920 virt/kvm/kvm_main.c trace_kvm_userspace_exit(vcpu->run->exit_reason, r);