run                19 arch/arm/include/asm/kvm_coproc.h int kvm_handle_cp10_id(struct kvm_vcpu *vcpu, struct kvm_run *run);
run                20 arch/arm/include/asm/kvm_coproc.h int kvm_handle_cp_0_13_access(struct kvm_vcpu *vcpu, struct kvm_run *run);
run                21 arch/arm/include/asm/kvm_coproc.h int kvm_handle_cp14_load_store(struct kvm_vcpu *vcpu, struct kvm_run *run);
run                22 arch/arm/include/asm/kvm_coproc.h int kvm_handle_cp14_32(struct kvm_vcpu *vcpu, struct kvm_run *run);
run                23 arch/arm/include/asm/kvm_coproc.h int kvm_handle_cp14_64(struct kvm_vcpu *vcpu, struct kvm_run *run);
run                24 arch/arm/include/asm/kvm_coproc.h int kvm_handle_cp15_32(struct kvm_vcpu *vcpu, struct kvm_run *run);
run                25 arch/arm/include/asm/kvm_coproc.h int kvm_handle_cp15_64(struct kvm_vcpu *vcpu, struct kvm_run *run);
run               287 arch/arm/include/asm/kvm_host.h int handle_exit(struct kvm_vcpu *vcpu, struct kvm_run *run,
run               290 arch/arm/include/asm/kvm_host.h static inline void handle_exit_early(struct kvm_vcpu *vcpu, struct kvm_run *run,
run                24 arch/arm/include/asm/kvm_mmio.h int kvm_handle_mmio_return(struct kvm_vcpu *vcpu, struct kvm_run *run);
run                25 arch/arm/include/asm/kvm_mmio.h int io_mem_abort(struct kvm_vcpu *vcpu, struct kvm_run *run,
run                60 arch/arm/include/asm/kvm_mmu.h int kvm_handle_guest_abort(struct kvm_vcpu *vcpu, struct kvm_run *run);
run                81 arch/arm/kvm/coproc.c int kvm_handle_cp10_id(struct kvm_vcpu *vcpu, struct kvm_run *run)
run                87 arch/arm/kvm/coproc.c int kvm_handle_cp_0_13_access(struct kvm_vcpu *vcpu, struct kvm_run *run)
run                97 arch/arm/kvm/coproc.c int kvm_handle_cp14_load_store(struct kvm_vcpu *vcpu, struct kvm_run *run)
run               629 arch/arm/kvm/coproc.c int kvm_handle_cp15_64(struct kvm_vcpu *vcpu, struct kvm_run *run)
run               641 arch/arm/kvm/coproc.c int kvm_handle_cp14_64(struct kvm_vcpu *vcpu, struct kvm_run *run)
run               694 arch/arm/kvm/coproc.c int kvm_handle_cp15_32(struct kvm_vcpu *vcpu, struct kvm_run *run)
run               705 arch/arm/kvm/coproc.c int kvm_handle_cp14_32(struct kvm_vcpu *vcpu, struct kvm_run *run)
run                19 arch/arm/kvm/handle_exit.c static int handle_hvc(struct kvm_vcpu *vcpu, struct kvm_run *run)
run                36 arch/arm/kvm/handle_exit.c static int handle_smc(struct kvm_vcpu *vcpu, struct kvm_run *run)
run                62 arch/arm/kvm/handle_exit.c static int kvm_handle_wfx(struct kvm_vcpu *vcpu, struct kvm_run *run)
run                80 arch/arm/kvm/handle_exit.c static int kvm_handle_unknown_ec(struct kvm_vcpu *vcpu, struct kvm_run *run)
run               118 arch/arm/kvm/handle_exit.c int handle_exit(struct kvm_vcpu *vcpu, struct kvm_run *run,
run               157 arch/arm/kvm/handle_exit.c 		return exit_handler(vcpu, run);
run               167 arch/arm/kvm/handle_exit.c 		run->exit_reason = KVM_EXIT_FAIL_ENTRY;
run               172 arch/arm/kvm/handle_exit.c 		run->exit_reason = KVM_EXIT_INTERNAL_ERROR;
run                30 arch/arm64/include/asm/kvm_coproc.h int kvm_handle_cp14_load_store(struct kvm_vcpu *vcpu, struct kvm_run *run);
run                31 arch/arm64/include/asm/kvm_coproc.h int kvm_handle_cp14_32(struct kvm_vcpu *vcpu, struct kvm_run *run);
run                32 arch/arm64/include/asm/kvm_coproc.h int kvm_handle_cp14_64(struct kvm_vcpu *vcpu, struct kvm_run *run);
run                33 arch/arm64/include/asm/kvm_coproc.h int kvm_handle_cp15_32(struct kvm_vcpu *vcpu, struct kvm_run *run);
run                34 arch/arm64/include/asm/kvm_coproc.h int kvm_handle_cp15_64(struct kvm_vcpu *vcpu, struct kvm_run *run);
run                35 arch/arm64/include/asm/kvm_coproc.h int kvm_handle_sys_reg(struct kvm_vcpu *vcpu, struct kvm_run *run);
run               475 arch/arm64/include/asm/kvm_host.h int handle_exit(struct kvm_vcpu *vcpu, struct kvm_run *run,
run               477 arch/arm64/include/asm/kvm_host.h void handle_exit_early(struct kvm_vcpu *vcpu, struct kvm_run *run,
run                23 arch/arm64/include/asm/kvm_mmio.h int kvm_handle_mmio_return(struct kvm_vcpu *vcpu, struct kvm_run *run);
run                24 arch/arm64/include/asm/kvm_mmio.h int io_mem_abort(struct kvm_vcpu *vcpu, struct kvm_run *run,
run               161 arch/arm64/include/asm/kvm_mmu.h int kvm_handle_guest_abort(struct kvm_vcpu *vcpu, struct kvm_run *run);
run                36 arch/arm64/kvm/handle_exit.c static int handle_hvc(struct kvm_vcpu *vcpu, struct kvm_run *run)
run                53 arch/arm64/kvm/handle_exit.c static int handle_smc(struct kvm_vcpu *vcpu, struct kvm_run *run)
run                72 arch/arm64/kvm/handle_exit.c static int handle_no_fpsimd(struct kvm_vcpu *vcpu, struct kvm_run *run)
run                90 arch/arm64/kvm/handle_exit.c static int kvm_handle_wfx(struct kvm_vcpu *vcpu, struct kvm_run *run)
run               120 arch/arm64/kvm/handle_exit.c static int kvm_handle_guest_debug(struct kvm_vcpu *vcpu, struct kvm_run *run)
run               125 arch/arm64/kvm/handle_exit.c 	run->exit_reason = KVM_EXIT_DEBUG;
run               126 arch/arm64/kvm/handle_exit.c 	run->debug.arch.hsr = hsr;
run               130 arch/arm64/kvm/handle_exit.c 		run->debug.arch.far = vcpu->arch.fault.far_el2;
run               147 arch/arm64/kvm/handle_exit.c static int kvm_handle_unknown_ec(struct kvm_vcpu *vcpu, struct kvm_run *run)
run               158 arch/arm64/kvm/handle_exit.c static int handle_sve(struct kvm_vcpu *vcpu, struct kvm_run *run)
run               181 arch/arm64/kvm/handle_exit.c static int kvm_handle_ptrauth(struct kvm_vcpu *vcpu, struct kvm_run *run)
run               226 arch/arm64/kvm/handle_exit.c static int handle_trap_exceptions(struct kvm_vcpu *vcpu, struct kvm_run *run)
run               241 arch/arm64/kvm/handle_exit.c 		handled = exit_handler(vcpu, run);
run               251 arch/arm64/kvm/handle_exit.c int handle_exit(struct kvm_vcpu *vcpu, struct kvm_run *run,
run               279 arch/arm64/kvm/handle_exit.c 		return handle_trap_exceptions(vcpu, run);
run               285 arch/arm64/kvm/handle_exit.c 		run->exit_reason = KVM_EXIT_FAIL_ENTRY;
run               292 arch/arm64/kvm/handle_exit.c 		run->exit_reason = KVM_EXIT_FAIL_ENTRY;
run               297 arch/arm64/kvm/handle_exit.c 		run->exit_reason = KVM_EXIT_INTERNAL_ERROR;
run               303 arch/arm64/kvm/handle_exit.c void handle_exit_early(struct kvm_vcpu *vcpu, struct kvm_run *run,
run              2027 arch/arm64/kvm/sys_regs.c int kvm_handle_cp14_load_store(struct kvm_vcpu *vcpu, struct kvm_run *run)
run              2206 arch/arm64/kvm/sys_regs.c int kvm_handle_cp15_64(struct kvm_vcpu *vcpu, struct kvm_run *run)
run              2217 arch/arm64/kvm/sys_regs.c int kvm_handle_cp15_32(struct kvm_vcpu *vcpu, struct kvm_run *run)
run              2228 arch/arm64/kvm/sys_regs.c int kvm_handle_cp14_64(struct kvm_vcpu *vcpu, struct kvm_run *run)
run              2235 arch/arm64/kvm/sys_regs.c int kvm_handle_cp14_32(struct kvm_vcpu *vcpu, struct kvm_run *run)
run              2287 arch/arm64/kvm/sys_regs.c int kvm_handle_sys_reg(struct kvm_vcpu *vcpu, struct kvm_run *run)
run               317 arch/mips/include/asm/kvm_host.h 	int (*vcpu_run)(struct kvm_run *run, struct kvm_vcpu *vcpu);
run               819 arch/mips/include/asm/kvm_host.h 	int (*vcpu_run)(struct kvm_run *run, struct kvm_vcpu *vcpu);
run               820 arch/mips/include/asm/kvm_host.h 	void (*vcpu_reenter)(struct kvm_run *run, struct kvm_vcpu *vcpu);
run               828 arch/mips/include/asm/kvm_host.h extern int kvm_mips_handle_exit(struct kvm_run *run, struct kvm_vcpu *vcpu);
run               875 arch/mips/include/asm/kvm_host.h 						     struct kvm_run *run,
run               982 arch/mips/include/asm/kvm_host.h 						   struct kvm_run *run,
run               989 arch/mips/include/asm/kvm_host.h 						      struct kvm_run *run,
run               994 arch/mips/include/asm/kvm_host.h 							 struct kvm_run *run,
run               999 arch/mips/include/asm/kvm_host.h 							struct kvm_run *run,
run              1004 arch/mips/include/asm/kvm_host.h 							 struct kvm_run *run,
run              1009 arch/mips/include/asm/kvm_host.h 							struct kvm_run *run,
run              1014 arch/mips/include/asm/kvm_host.h 						     struct kvm_run *run,
run              1019 arch/mips/include/asm/kvm_host.h 						      struct kvm_run *run,
run              1024 arch/mips/include/asm/kvm_host.h 						struct kvm_run *run,
run              1029 arch/mips/include/asm/kvm_host.h 						     struct kvm_run *run,
run              1034 arch/mips/include/asm/kvm_host.h 						     struct kvm_run *run,
run              1039 arch/mips/include/asm/kvm_host.h 						       struct kvm_run *run,
run              1044 arch/mips/include/asm/kvm_host.h 							 struct kvm_run *run,
run              1049 arch/mips/include/asm/kvm_host.h 						      struct kvm_run *run,
run              1054 arch/mips/include/asm/kvm_host.h 							 struct kvm_run *run,
run              1058 arch/mips/include/asm/kvm_host.h 							 struct kvm_run *run);
run              1087 arch/mips/include/asm/kvm_host.h 					       struct kvm_run *run,
run              1093 arch/mips/include/asm/kvm_host.h 					     struct kvm_run *run,
run              1098 arch/mips/include/asm/kvm_host.h 					   struct kvm_run *run,
run              1102 arch/mips/include/asm/kvm_host.h 					     struct kvm_run *run,
run              1106 arch/mips/include/asm/kvm_host.h 					    struct kvm_run *run,
run               987 arch/mips/kvm/emulate.c 			vcpu->run->exit_reason = KVM_EXIT_IRQ_WINDOW_OPEN;
run              1265 arch/mips/kvm/emulate.c 					   struct kvm_run *run,
run              1600 arch/mips/kvm/emulate.c 					     struct kvm_run *run,
run              1605 arch/mips/kvm/emulate.c 	void *data = run->mmio.data;
run              1619 arch/mips/kvm/emulate.c 	run->mmio.phys_addr = kvm_mips_callbacks->gva_to_gpa(
run              1621 arch/mips/kvm/emulate.c 	if (run->mmio.phys_addr == KVM_INVALID_ADDR)
run              1627 arch/mips/kvm/emulate.c 		run->mmio.len = 8;
run              1637 arch/mips/kvm/emulate.c 		run->mmio.len = 4;
run              1646 arch/mips/kvm/emulate.c 		run->mmio.len = 2;
run              1655 arch/mips/kvm/emulate.c 		run->mmio.len = 1;
run              1669 arch/mips/kvm/emulate.c 	run->mmio.is_write = 1;
run              1681 arch/mips/kvm/emulate.c 					    u32 cause, struct kvm_run *run,
run              1705 arch/mips/kvm/emulate.c 	run->mmio.phys_addr = kvm_mips_callbacks->gva_to_gpa(
run              1707 arch/mips/kvm/emulate.c 	if (run->mmio.phys_addr == KVM_INVALID_ADDR)
run              1714 arch/mips/kvm/emulate.c 		run->mmio.len = 8;
run              1722 arch/mips/kvm/emulate.c 		run->mmio.len = 4;
run              1729 arch/mips/kvm/emulate.c 		run->mmio.len = 2;
run              1736 arch/mips/kvm/emulate.c 		run->mmio.len = 1;
run              1746 arch/mips/kvm/emulate.c 	run->mmio.is_write = 0;
run              1755 arch/mips/kvm/emulate.c 						     struct kvm_run *run,
run              1783 arch/mips/kvm/emulate.c 			kvm_mips_emulate_tlbmiss_ld(cause, NULL, run, vcpu);
run              1789 arch/mips/kvm/emulate.c 			kvm_mips_emulate_tlbinv_ld(cause, NULL, run, vcpu);
run              1799 arch/mips/kvm/emulate.c 					     struct kvm_run *run,
run              1889 arch/mips/kvm/emulate.c 					     curr_pc, va, run, vcpu, cause);
run              1902 arch/mips/kvm/emulate.c 					     curr_pc, va, run, vcpu, cause);
run              1906 arch/mips/kvm/emulate.c 					     curr_pc, va, run, vcpu, cause);
run              1932 arch/mips/kvm/emulate.c 					    struct kvm_run *run,
run              1948 arch/mips/kvm/emulate.c 		er = kvm_mips_emulate_CP0(inst, opc, cause, run, vcpu);
run              1955 arch/mips/kvm/emulate.c 		er = kvm_mips_emulate_cache(inst, opc, cause, run, vcpu);
run              1963 arch/mips/kvm/emulate.c 			er = kvm_mips_emulate_cache(inst, opc, cause, run,
run              2003 arch/mips/kvm/emulate.c 					       struct kvm_run *run,
run              2038 arch/mips/kvm/emulate.c 						  struct kvm_run *run,
run              2082 arch/mips/kvm/emulate.c 						 struct kvm_run *run,
run              2124 arch/mips/kvm/emulate.c 						  struct kvm_run *run,
run              2166 arch/mips/kvm/emulate.c 						 struct kvm_run *run,
run              2207 arch/mips/kvm/emulate.c 					      struct kvm_run *run,
run              2247 arch/mips/kvm/emulate.c 					       struct kvm_run *run,
run              2276 arch/mips/kvm/emulate.c 					      struct kvm_run *run,
run              2311 arch/mips/kvm/emulate.c 					      struct kvm_run *run,
run              2346 arch/mips/kvm/emulate.c 						struct kvm_run *run,
run              2381 arch/mips/kvm/emulate.c 						  struct kvm_run *run,
run              2416 arch/mips/kvm/emulate.c 					       struct kvm_run *run,
run              2451 arch/mips/kvm/emulate.c 						  struct kvm_run *run,
run              2485 arch/mips/kvm/emulate.c 					 struct kvm_run *run,
run              2574 arch/mips/kvm/emulate.c 	return kvm_mips_emulate_ri_exc(cause, opc, run, vcpu);
run              2578 arch/mips/kvm/emulate.c 						  struct kvm_run *run)
run              2583 arch/mips/kvm/emulate.c 	if (run->mmio.len > sizeof(*gpr)) {
run              2584 arch/mips/kvm/emulate.c 		kvm_err("Bad MMIO length: %d", run->mmio.len);
run              2592 arch/mips/kvm/emulate.c 	switch (run->mmio.len) {
run              2594 arch/mips/kvm/emulate.c 		*gpr = *(s64 *)run->mmio.data;
run              2599 arch/mips/kvm/emulate.c 			*gpr = *(s32 *)run->mmio.data;
run              2601 arch/mips/kvm/emulate.c 			*gpr = *(u32 *)run->mmio.data;
run              2606 arch/mips/kvm/emulate.c 			*gpr = *(s16 *) run->mmio.data;
run              2608 arch/mips/kvm/emulate.c 			*gpr = *(u16 *)run->mmio.data;
run              2613 arch/mips/kvm/emulate.c 			*gpr = *(s8 *) run->mmio.data;
run              2615 arch/mips/kvm/emulate.c 			*gpr = *(u8 *) run->mmio.data;
run              2625 arch/mips/kvm/emulate.c 						  struct kvm_run *run,
run              2663 arch/mips/kvm/emulate.c 					       struct kvm_run *run,
run              2745 arch/mips/kvm/emulate.c 		kvm_mips_emulate_exc(cause, opc, run, vcpu);
run              2759 arch/mips/kvm/emulate.c 					      struct kvm_run *run,
run              2783 arch/mips/kvm/emulate.c 			er = kvm_mips_emulate_tlbmiss_ld(cause, opc, run, vcpu);
run              2785 arch/mips/kvm/emulate.c 			er = kvm_mips_emulate_tlbmiss_st(cause, opc, run, vcpu);
run              2800 arch/mips/kvm/emulate.c 				er = kvm_mips_emulate_tlbinv_ld(cause, opc, run,
run              2803 arch/mips/kvm/emulate.c 				er = kvm_mips_emulate_tlbinv_st(cause, opc, run,
run               634 arch/mips/kvm/entry.c 	UASM_i_LW(&p, S0, offsetof(struct kvm_vcpu, run), S1);
run               435 arch/mips/kvm/mips.c int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu, struct kvm_run *run)
run               445 arch/mips/kvm/mips.c 			kvm_mips_complete_mmio_load(vcpu, run);
run               449 arch/mips/kvm/mips.c 	if (run->immediate_exit)
run               466 arch/mips/kvm/mips.c 	r = kvm_mips_callbacks->vcpu_run(run, vcpu);
run              1281 arch/mips/kvm/mips.c int kvm_mips_handle_exit(struct kvm_run *run, struct kvm_vcpu *vcpu)
run              1298 arch/mips/kvm/mips.c 	run->exit_reason = KVM_EXIT_UNKNOWN;
run              1299 arch/mips/kvm/mips.c 	run->ready_for_interrupt_injection = 1;
run              1310 arch/mips/kvm/mips.c 			cause, opc, run, vcpu);
run              1319 arch/mips/kvm/mips.c 		er = kvm_mips_check_privilege(cause, opc, run, vcpu);
run              1323 arch/mips/kvm/mips.c 			run->exit_reason = KVM_EXIT_INTERNAL_ERROR;
run              1347 arch/mips/kvm/mips.c 		if (run->exit_reason == KVM_EXIT_IRQ_WINDOW_OPEN)
run              1432 arch/mips/kvm/mips.c 		run->exit_reason = KVM_EXIT_INTERNAL_ERROR;
run              1450 arch/mips/kvm/mips.c 			run->exit_reason = KVM_EXIT_INTR;
run              1468 arch/mips/kvm/mips.c 		kvm_mips_callbacks->vcpu_reenter(run, vcpu);
run                63 arch/mips/kvm/trap_emul.c 	vcpu->run->exit_reason = KVM_EXIT_INTERNAL_ERROR;
run                70 arch/mips/kvm/trap_emul.c 	struct kvm_run *run = vcpu->run;
run                84 arch/mips/kvm/trap_emul.c 			er = kvm_mips_emulate_fpu_exc(cause, opc, run, vcpu);
run                91 arch/mips/kvm/trap_emul.c 		er = kvm_mips_emulate_inst(cause, opc, run, vcpu);
run               100 arch/mips/kvm/trap_emul.c 		run->exit_reason = KVM_EXIT_INTERNAL_ERROR;
run               105 arch/mips/kvm/trap_emul.c 		run->exit_reason = KVM_EXIT_INTR;
run               119 arch/mips/kvm/trap_emul.c static int kvm_mips_bad_load(u32 cause, u32 *opc, struct kvm_run *run,
run               128 arch/mips/kvm/trap_emul.c 		run->exit_reason = KVM_EXIT_INTERNAL_ERROR;
run               137 arch/mips/kvm/trap_emul.c 		run->exit_reason = KVM_EXIT_INTERNAL_ERROR;
run               142 arch/mips/kvm/trap_emul.c 	er = kvm_mips_emulate_load(inst, cause, run, vcpu);
run               145 arch/mips/kvm/trap_emul.c 		run->exit_reason = KVM_EXIT_INTERNAL_ERROR;
run               147 arch/mips/kvm/trap_emul.c 		run->exit_reason = KVM_EXIT_MMIO;
run               152 arch/mips/kvm/trap_emul.c static int kvm_mips_bad_store(u32 cause, u32 *opc, struct kvm_run *run,
run               164 arch/mips/kvm/trap_emul.c 		run->exit_reason = KVM_EXIT_INTERNAL_ERROR;
run               169 arch/mips/kvm/trap_emul.c 	er = kvm_mips_emulate_store(inst, cause, run, vcpu);
run               172 arch/mips/kvm/trap_emul.c 		run->exit_reason = KVM_EXIT_INTERNAL_ERROR;
run               174 arch/mips/kvm/trap_emul.c 		run->exit_reason = KVM_EXIT_MMIO;
run               179 arch/mips/kvm/trap_emul.c static int kvm_mips_bad_access(u32 cause, u32 *opc, struct kvm_run *run,
run               183 arch/mips/kvm/trap_emul.c 		return kvm_mips_bad_store(cause, opc, run, vcpu);
run               185 arch/mips/kvm/trap_emul.c 		return kvm_mips_bad_load(cause, opc, run, vcpu);
run               191 arch/mips/kvm/trap_emul.c 	struct kvm_run *run = vcpu->run;
run               215 arch/mips/kvm/trap_emul.c 			run->exit_reason = KVM_EXIT_INTERNAL_ERROR;
run               220 arch/mips/kvm/trap_emul.c 			run->exit_reason = KVM_EXIT_INTERNAL_ERROR;
run               229 arch/mips/kvm/trap_emul.c 			kvm_mips_emulate_tlbmod(cause, opc, run, vcpu);
run               236 arch/mips/kvm/trap_emul.c 			return kvm_mips_bad_store(cause, opc, run, vcpu);
run               241 arch/mips/kvm/trap_emul.c 			return kvm_mips_bad_store(cause, opc, run, vcpu);
run               245 arch/mips/kvm/trap_emul.c 		return kvm_mips_bad_store(cause, opc, run, vcpu);
run               251 arch/mips/kvm/trap_emul.c 	struct kvm_run *run = vcpu->run;
run               261 arch/mips/kvm/trap_emul.c 			run->exit_reason = KVM_EXIT_INTERNAL_ERROR;
run               279 arch/mips/kvm/trap_emul.c 		er = kvm_mips_handle_tlbmiss(cause, opc, run, vcpu, store);
run               283 arch/mips/kvm/trap_emul.c 			run->exit_reason = KVM_EXIT_INTERNAL_ERROR;
run               292 arch/mips/kvm/trap_emul.c 			ret = kvm_mips_bad_access(cause, opc, run, vcpu, store);
run               299 arch/mips/kvm/trap_emul.c 		ret = kvm_mips_bad_access(cause, opc, run, vcpu, store);
run               305 arch/mips/kvm/trap_emul.c 		run->exit_reason = KVM_EXIT_INTERNAL_ERROR;
run               323 arch/mips/kvm/trap_emul.c 	struct kvm_run *run = vcpu->run;
run               331 arch/mips/kvm/trap_emul.c 		ret = kvm_mips_bad_store(cause, opc, run, vcpu);
run               335 arch/mips/kvm/trap_emul.c 		run->exit_reason = KVM_EXIT_INTERNAL_ERROR;
run               343 arch/mips/kvm/trap_emul.c 	struct kvm_run *run = vcpu->run;
run               350 arch/mips/kvm/trap_emul.c 		ret = kvm_mips_bad_load(cause, opc, run, vcpu);
run               354 arch/mips/kvm/trap_emul.c 		run->exit_reason = KVM_EXIT_INTERNAL_ERROR;
run               362 arch/mips/kvm/trap_emul.c 	struct kvm_run *run = vcpu->run;
run               368 arch/mips/kvm/trap_emul.c 	er = kvm_mips_emulate_syscall(cause, opc, run, vcpu);
run               372 arch/mips/kvm/trap_emul.c 		run->exit_reason = KVM_EXIT_INTERNAL_ERROR;
run               380 arch/mips/kvm/trap_emul.c 	struct kvm_run *run = vcpu->run;
run               386 arch/mips/kvm/trap_emul.c 	er = kvm_mips_handle_ri(cause, opc, run, vcpu);
run               390 arch/mips/kvm/trap_emul.c 		run->exit_reason = KVM_EXIT_INTERNAL_ERROR;
run               398 arch/mips/kvm/trap_emul.c 	struct kvm_run *run = vcpu->run;
run               404 arch/mips/kvm/trap_emul.c 	er = kvm_mips_emulate_bp_exc(cause, opc, run, vcpu);
run               408 arch/mips/kvm/trap_emul.c 		run->exit_reason = KVM_EXIT_INTERNAL_ERROR;
run               416 arch/mips/kvm/trap_emul.c 	struct kvm_run *run = vcpu->run;
run               422 arch/mips/kvm/trap_emul.c 	er = kvm_mips_emulate_trap_exc(cause, opc, run, vcpu);
run               426 arch/mips/kvm/trap_emul.c 		run->exit_reason = KVM_EXIT_INTERNAL_ERROR;
run               434 arch/mips/kvm/trap_emul.c 	struct kvm_run *run = vcpu->run;
run               440 arch/mips/kvm/trap_emul.c 	er = kvm_mips_emulate_msafpe_exc(cause, opc, run, vcpu);
run               444 arch/mips/kvm/trap_emul.c 		run->exit_reason = KVM_EXIT_INTERNAL_ERROR;
run               452 arch/mips/kvm/trap_emul.c 	struct kvm_run *run = vcpu->run;
run               458 arch/mips/kvm/trap_emul.c 	er = kvm_mips_emulate_fpe_exc(cause, opc, run, vcpu);
run               462 arch/mips/kvm/trap_emul.c 		run->exit_reason = KVM_EXIT_INTERNAL_ERROR;
run               477 arch/mips/kvm/trap_emul.c 	struct kvm_run *run = vcpu->run;
run               489 arch/mips/kvm/trap_emul.c 		er = kvm_mips_emulate_ri_exc(cause, opc, run, vcpu);
run               492 arch/mips/kvm/trap_emul.c 		er = kvm_mips_emulate_msadis_exc(cause, opc, run, vcpu);
run               505 arch/mips/kvm/trap_emul.c 		run->exit_reason = KVM_EXIT_INTERNAL_ERROR;
run              1182 arch/mips/kvm/trap_emul.c static void kvm_trap_emul_vcpu_reenter(struct kvm_run *run,
run              1226 arch/mips/kvm/trap_emul.c static int kvm_trap_emul_vcpu_run(struct kvm_run *run, struct kvm_vcpu *vcpu)
run              1235 arch/mips/kvm/trap_emul.c 	kvm_trap_emul_vcpu_reenter(run, vcpu);
run              1253 arch/mips/kvm/trap_emul.c 	r = vcpu->arch.vcpu_run(run, vcpu);
run               861 arch/mips/kvm/vz.c 	vcpu->run->exit_reason = KVM_EXIT_INTERNAL_ERROR;
run               902 arch/mips/kvm/vz.c 					      struct kvm_run *run,
run              1065 arch/mips/kvm/vz.c 					       struct kvm_run *run,
run              1137 arch/mips/kvm/vz.c 	struct kvm_run *run = vcpu->run;
run              1153 arch/mips/kvm/vz.c 		er = kvm_vz_gpsi_cop0(inst, opc, cause, run, vcpu);
run              1158 arch/mips/kvm/vz.c 		er = kvm_vz_gpsi_cache(inst, opc, cause, run, vcpu);
run              1166 arch/mips/kvm/vz.c 			er = kvm_vz_gpsi_cache(inst, opc, cause, run, vcpu);
run              1453 arch/mips/kvm/vz.c 		vcpu->run->exit_reason = KVM_EXIT_INTERNAL_ERROR;
run              1468 arch/mips/kvm/vz.c 	struct kvm_run *run = vcpu->run;
run              1496 arch/mips/kvm/vz.c 		run->exit_reason = KVM_EXIT_INTERNAL_ERROR;
run              1515 arch/mips/kvm/vz.c 	struct kvm_run *run = vcpu->run;
run              1527 arch/mips/kvm/vz.c 		run->exit_reason = KVM_EXIT_INTERNAL_ERROR;
run              1538 arch/mips/kvm/vz.c 	struct kvm_run *run = vcpu->run;
run              1549 arch/mips/kvm/vz.c 			run->exit_reason = KVM_EXIT_INTERNAL_ERROR;
run              1558 arch/mips/kvm/vz.c 			run->exit_reason = KVM_EXIT_INTERNAL_ERROR;
run              1563 arch/mips/kvm/vz.c 		er = kvm_mips_emulate_load(inst, cause, run, vcpu);
run              1567 arch/mips/kvm/vz.c 			run->exit_reason = KVM_EXIT_INTERNAL_ERROR;
run              1574 arch/mips/kvm/vz.c 		run->exit_reason = KVM_EXIT_MMIO;
run              1577 arch/mips/kvm/vz.c 		run->exit_reason = KVM_EXIT_INTERNAL_ERROR;
run              1585 arch/mips/kvm/vz.c 	struct kvm_run *run = vcpu->run;
run              1605 arch/mips/kvm/vz.c 			run->exit_reason = KVM_EXIT_INTERNAL_ERROR;
run              1610 arch/mips/kvm/vz.c 		er = kvm_mips_emulate_store(inst, cause, run, vcpu);
run              1614 arch/mips/kvm/vz.c 			run->exit_reason = KVM_EXIT_INTERNAL_ERROR;
run              1621 arch/mips/kvm/vz.c 		run->exit_reason = KVM_EXIT_MMIO;
run              1624 arch/mips/kvm/vz.c 		run->exit_reason = KVM_EXIT_INTERNAL_ERROR;
run              3132 arch/mips/kvm/vz.c static void kvm_vz_vcpu_reenter(struct kvm_run *run, struct kvm_vcpu *vcpu)
run              3148 arch/mips/kvm/vz.c static int kvm_vz_vcpu_run(struct kvm_run *run, struct kvm_vcpu *vcpu)
run              3161 arch/mips/kvm/vz.c 	r = vcpu->arch.vcpu_run(run, vcpu);
run               158 arch/powerpc/include/asm/kvm_book3s.h extern int kvmppc_book3s_hv_page_fault(struct kvm_run *run,
run               163 arch/powerpc/include/asm/kvm_book3s.h extern int kvmppc_hv_emulate_mmio(struct kvm_run *run, struct kvm_vcpu *vcpu,
run               177 arch/powerpc/include/asm/kvm_book3s.h extern int kvmppc_book3s_radix_page_fault(struct kvm_run *run,
run               237 arch/powerpc/include/asm/kvm_book3s.h extern int kvmppc_emulate_paired_single(struct kvm_run *run, struct kvm_vcpu *vcpu);
run               308 arch/powerpc/include/asm/kvm_book3s.h long int kvmhv_nested_page_fault(struct kvm_run *run, struct kvm_vcpu *vcpu);
run                66 arch/powerpc/include/asm/kvm_ppc.h extern int kvmppc_handle_load(struct kvm_run *run, struct kvm_vcpu *vcpu,
run                69 arch/powerpc/include/asm/kvm_ppc.h extern int kvmppc_handle_loads(struct kvm_run *run, struct kvm_vcpu *vcpu,
run                72 arch/powerpc/include/asm/kvm_ppc.h extern int kvmppc_handle_vsx_load(struct kvm_run *run, struct kvm_vcpu *vcpu,
run                75 arch/powerpc/include/asm/kvm_ppc.h extern int kvmppc_handle_vmx_load(struct kvm_run *run, struct kvm_vcpu *vcpu,
run                77 arch/powerpc/include/asm/kvm_ppc.h extern int kvmppc_handle_vmx_store(struct kvm_run *run, struct kvm_vcpu *vcpu,
run                79 arch/powerpc/include/asm/kvm_ppc.h extern int kvmppc_handle_store(struct kvm_run *run, struct kvm_vcpu *vcpu,
run                82 arch/powerpc/include/asm/kvm_ppc.h extern int kvmppc_handle_vsx_store(struct kvm_run *run, struct kvm_vcpu *vcpu,
run                93 arch/powerpc/include/asm/kvm_ppc.h extern int kvmppc_emulate_instruction(struct kvm_run *run,
run                96 arch/powerpc/include/asm/kvm_ppc.h extern int kvmppc_emulate_mmio(struct kvm_run *run, struct kvm_vcpu *vcpu);
run               275 arch/powerpc/include/asm/kvm_ppc.h 	int (*vcpu_run)(struct kvm_run *run, struct kvm_vcpu *vcpu);
run               302 arch/powerpc/include/asm/kvm_ppc.h 	int (*emulate_op)(struct kvm_run *run, struct kvm_vcpu *vcpu,
run                20 arch/powerpc/kvm/book3s.h extern int kvmppc_core_emulate_op_pr(struct kvm_run *run, struct kvm_vcpu *vcpu,
run               428 arch/powerpc/kvm/book3s_64_mmu_hv.c int kvmppc_hv_emulate_mmio(struct kvm_run *run, struct kvm_vcpu *vcpu,
run               488 arch/powerpc/kvm/book3s_64_mmu_hv.c 	return kvmppc_emulate_mmio(run, vcpu);
run               491 arch/powerpc/kvm/book3s_64_mmu_hv.c int kvmppc_book3s_hv_page_fault(struct kvm_run *run, struct kvm_vcpu *vcpu,
run               513 arch/powerpc/kvm/book3s_64_mmu_hv.c 		return kvmppc_book3s_radix_page_fault(run, vcpu, ea, dsisr);
run               533 arch/powerpc/kvm/book3s_64_mmu_hv.c 			return kvmppc_hv_emulate_mmio(run, vcpu, gpa, ea,
run               569 arch/powerpc/kvm/book3s_64_mmu_hv.c 		return kvmppc_hv_emulate_mmio(run, vcpu, gpa, ea,
run               889 arch/powerpc/kvm/book3s_64_mmu_radix.c int kvmppc_book3s_radix_page_fault(struct kvm_run *run, struct kvm_vcpu *vcpu,
run               932 arch/powerpc/kvm/book3s_64_mmu_radix.c 		return kvmppc_hv_emulate_mmio(run, vcpu, gpa, ea, writing);
run               238 arch/powerpc/kvm/book3s_emulate.c int kvmppc_core_emulate_op_pr(struct kvm_run *run, struct kvm_vcpu *vcpu,
run               374 arch/powerpc/kvm/book3s_emulate.c 			run->papr_hcall.nr = cmd;
run               377 arch/powerpc/kvm/book3s_emulate.c 				run->papr_hcall.args[i] = gpr;
run               380 arch/powerpc/kvm/book3s_emulate.c 			run->exit_reason = KVM_EXIT_PAPR_HCALL;
run               632 arch/powerpc/kvm/book3s_emulate.c 		emulated = kvmppc_emulate_paired_single(run, vcpu);
run              1133 arch/powerpc/kvm/book3s_hv.c static int kvmppc_emulate_debug_inst(struct kvm_run *run,
run              1148 arch/powerpc/kvm/book3s_hv.c 		run->exit_reason = KVM_EXIT_DEBUG;
run              1149 arch/powerpc/kvm/book3s_hv.c 		run->debug.arch.address = kvmppc_get_pc(vcpu);
run              1250 arch/powerpc/kvm/book3s_hv.c static int kvmppc_handle_exit_hv(struct kvm_run *run, struct kvm_vcpu *vcpu,
run              1271 arch/powerpc/kvm/book3s_hv.c 		run->exit_reason = KVM_EXIT_INTERNAL_ERROR;
run              1272 arch/powerpc/kvm/book3s_hv.c 		run->hw.hardware_exit_reason = vcpu->arch.trap;
run              1275 arch/powerpc/kvm/book3s_hv.c 	run->exit_reason = KVM_EXIT_UNKNOWN;
run              1276 arch/powerpc/kvm/book3s_hv.c 	run->ready_for_interrupt_injection = 1;
run              1313 arch/powerpc/kvm/book3s_hv.c 		run->exit_reason = KVM_EXIT_NMI;
run              1314 arch/powerpc/kvm/book3s_hv.c 		run->hw.hardware_exit_reason = vcpu->arch.trap;
run              1316 arch/powerpc/kvm/book3s_hv.c 		run->flags &= ~KVM_RUN_PPC_NMI_DISP_MASK;
run              1319 arch/powerpc/kvm/book3s_hv.c 			run->flags |= KVM_RUN_PPC_NMI_DISP_FULLY_RECOV;
run              1321 arch/powerpc/kvm/book3s_hv.c 			run->flags |= KVM_RUN_PPC_NMI_DISP_NOT_RECOV;
run              1348 arch/powerpc/kvm/book3s_hv.c 		run->papr_hcall.nr = kvmppc_get_gpr(vcpu, 3);
run              1350 arch/powerpc/kvm/book3s_hv.c 			run->papr_hcall.args[i] = kvmppc_get_gpr(vcpu, 4 + i);
run              1351 arch/powerpc/kvm/book3s_hv.c 		run->exit_reason = KVM_EXIT_PAPR_HCALL;
run              1387 arch/powerpc/kvm/book3s_hv.c 			r = kvmppc_emulate_debug_inst(run, vcpu);
run              1431 arch/powerpc/kvm/book3s_hv.c 		run->hw.hardware_exit_reason = vcpu->arch.trap;
run              1439 arch/powerpc/kvm/book3s_hv.c static int kvmppc_handle_nested_exit(struct kvm_run *run, struct kvm_vcpu *vcpu)
run              1497 arch/powerpc/kvm/book3s_hv.c 		r = kvmhv_nested_page_fault(run, vcpu);
run              1507 arch/powerpc/kvm/book3s_hv.c 		r = kvmhv_nested_page_fault(run, vcpu);
run              4215 arch/powerpc/kvm/book3s_hv.c static int kvmppc_vcpu_run_hv(struct kvm_run *run, struct kvm_vcpu *vcpu)
run              4225 arch/powerpc/kvm/book3s_hv.c 		run->exit_reason = KVM_EXIT_INTERNAL_ERROR;
run              4239 arch/powerpc/kvm/book3s_hv.c 			run->exit_reason = KVM_EXIT_FAIL_ENTRY;
run              4240 arch/powerpc/kvm/book3s_hv.c 			run->fail_entry.hardware_entry_failure_reason = 0;
run              4265 arch/powerpc/kvm/book3s_hv.c 		run->exit_reason = KVM_EXIT_INTR;
run              4300 arch/powerpc/kvm/book3s_hv.c 			r = kvmhv_run_single_vcpu(run, vcpu, ~(u64)0,
run              4303 arch/powerpc/kvm/book3s_hv.c 			r = kvmppc_run_vcpu(run, vcpu);
run              4305 arch/powerpc/kvm/book3s_hv.c 		if (run->exit_reason == KVM_EXIT_PAPR_HCALL &&
run              4313 arch/powerpc/kvm/book3s_hv.c 			r = kvmppc_book3s_hv_page_fault(run, vcpu,
run              4967 arch/powerpc/kvm/book3s_hv.c static int kvmppc_core_emulate_op_hv(struct kvm_run *run, struct kvm_vcpu *vcpu,
run              1260 arch/powerpc/kvm/book3s_hv_nested.c static long int __kvmhv_nested_page_fault(struct kvm_run *run,
run              1344 arch/powerpc/kvm/book3s_hv_nested.c 		return kvmppc_hv_emulate_mmio(run, vcpu, gpa, ea, writing);
run              1431 arch/powerpc/kvm/book3s_hv_nested.c long int kvmhv_nested_page_fault(struct kvm_run *run, struct kvm_vcpu *vcpu)
run              1437 arch/powerpc/kvm/book3s_hv_nested.c 	ret = __kvmhv_nested_page_fault(run, vcpu, gp);
run               172 arch/powerpc/kvm/book3s_paired_singles.c static int kvmppc_emulate_fpr_load(struct kvm_run *run, struct kvm_vcpu *vcpu,
run               191 arch/powerpc/kvm/book3s_paired_singles.c 		emulated = kvmppc_handle_load(run, vcpu, KVM_MMIO_REG_FPR | rs,
run               216 arch/powerpc/kvm/book3s_paired_singles.c static int kvmppc_emulate_fpr_store(struct kvm_run *run, struct kvm_vcpu *vcpu,
run               251 arch/powerpc/kvm/book3s_paired_singles.c 		emulated = kvmppc_handle_store(run, vcpu, val, len, 1);
run               262 arch/powerpc/kvm/book3s_paired_singles.c static int kvmppc_emulate_psq_load(struct kvm_run *run, struct kvm_vcpu *vcpu,
run               282 arch/powerpc/kvm/book3s_paired_singles.c 		emulated = kvmppc_handle_load(run, vcpu, KVM_MMIO_REG_FPR | rs,
run               287 arch/powerpc/kvm/book3s_paired_singles.c 		emulated = kvmppc_handle_load(run, vcpu, KVM_MMIO_REG_FQPR | rs,
run               305 arch/powerpc/kvm/book3s_paired_singles.c static int kvmppc_emulate_psq_store(struct kvm_run *run, struct kvm_vcpu *vcpu,
run               321 arch/powerpc/kvm/book3s_paired_singles.c 		emulated = kvmppc_handle_store(run, vcpu, tmp[0], 4, 1);
run               324 arch/powerpc/kvm/book3s_paired_singles.c 		emulated = kvmppc_handle_store(run, vcpu, val, 8, 1);
run               621 arch/powerpc/kvm/book3s_paired_singles.c int kvmppc_emulate_paired_single(struct kvm_run *run, struct kvm_vcpu *vcpu)
run               683 arch/powerpc/kvm/book3s_paired_singles.c 		emulated = kvmppc_emulate_psq_load(run, vcpu, ax_rd, addr, w, i);
run               693 arch/powerpc/kvm/book3s_paired_singles.c 		emulated = kvmppc_emulate_psq_load(run, vcpu, ax_rd, addr, w, i);
run               706 arch/powerpc/kvm/book3s_paired_singles.c 		emulated = kvmppc_emulate_psq_store(run, vcpu, ax_rd, addr, w, i);
run               716 arch/powerpc/kvm/book3s_paired_singles.c 		emulated = kvmppc_emulate_psq_store(run, vcpu, ax_rd, addr, w, i);
run               736 arch/powerpc/kvm/book3s_paired_singles.c 			emulated = kvmppc_emulate_psq_load(run, vcpu, ax_rd, addr, w, i);
run               750 arch/powerpc/kvm/book3s_paired_singles.c 			emulated = kvmppc_emulate_psq_load(run, vcpu, ax_rd, addr, w, i);
run               827 arch/powerpc/kvm/book3s_paired_singles.c 			emulated = kvmppc_emulate_psq_store(run, vcpu, ax_rd, addr, w, i);
run               837 arch/powerpc/kvm/book3s_paired_singles.c 			emulated = kvmppc_emulate_psq_store(run, vcpu, ax_rd, addr, w, i);
run               925 arch/powerpc/kvm/book3s_paired_singles.c 		emulated = kvmppc_emulate_fpr_load(run, vcpu, ax_rd, addr,
run               933 arch/powerpc/kvm/book3s_paired_singles.c 		emulated = kvmppc_emulate_fpr_load(run, vcpu, ax_rd, addr,
run               944 arch/powerpc/kvm/book3s_paired_singles.c 		emulated = kvmppc_emulate_fpr_load(run, vcpu, ax_rd, addr,
run               952 arch/powerpc/kvm/book3s_paired_singles.c 		emulated = kvmppc_emulate_fpr_load(run, vcpu, ax_rd, addr,
run               963 arch/powerpc/kvm/book3s_paired_singles.c 		emulated = kvmppc_emulate_fpr_store(run, vcpu, ax_rd, addr,
run               971 arch/powerpc/kvm/book3s_paired_singles.c 		emulated = kvmppc_emulate_fpr_store(run, vcpu, ax_rd, addr,
run               982 arch/powerpc/kvm/book3s_paired_singles.c 		emulated = kvmppc_emulate_fpr_store(run, vcpu, ax_rd, addr,
run               990 arch/powerpc/kvm/book3s_paired_singles.c 		emulated = kvmppc_emulate_fpr_store(run, vcpu, ax_rd, addr,
run              1004 arch/powerpc/kvm/book3s_paired_singles.c 			emulated = kvmppc_emulate_fpr_load(run, vcpu, ax_rd,
run              1013 arch/powerpc/kvm/book3s_paired_singles.c 			emulated = kvmppc_emulate_fpr_load(run, vcpu, ax_rd,
run              1025 arch/powerpc/kvm/book3s_paired_singles.c 			emulated = kvmppc_emulate_fpr_load(run, vcpu, ax_rd,
run              1034 arch/powerpc/kvm/book3s_paired_singles.c 			emulated = kvmppc_emulate_fpr_load(run, vcpu, ax_rd,
run              1046 arch/powerpc/kvm/book3s_paired_singles.c 			emulated = kvmppc_emulate_fpr_store(run, vcpu, ax_rd,
run              1055 arch/powerpc/kvm/book3s_paired_singles.c 			emulated = kvmppc_emulate_fpr_store(run, vcpu, ax_rd,
run              1067 arch/powerpc/kvm/book3s_paired_singles.c 			emulated = kvmppc_emulate_fpr_store(run, vcpu, ax_rd,
run              1076 arch/powerpc/kvm/book3s_paired_singles.c 			emulated = kvmppc_emulate_fpr_store(run, vcpu, ax_rd,
run              1088 arch/powerpc/kvm/book3s_paired_singles.c 			emulated = kvmppc_emulate_fpr_store(run, vcpu, ax_rd,
run               667 arch/powerpc/kvm/book3s_pr.c int kvmppc_handle_pagefault(struct kvm_run *run, struct kvm_vcpu *vcpu,
run               762 arch/powerpc/kvm/book3s_pr.c 			run->exit_reason = KVM_EXIT_INTERNAL_ERROR;
run               775 arch/powerpc/kvm/book3s_pr.c 		r = kvmppc_emulate_mmio(run, vcpu);
run               959 arch/powerpc/kvm/book3s_pr.c 		er = kvmppc_emulate_instruction(vcpu->run, vcpu);
run              1056 arch/powerpc/kvm/book3s_pr.c static int kvmppc_exit_pr_progint(struct kvm_run *run, struct kvm_vcpu *vcpu,
run              1091 arch/powerpc/kvm/book3s_pr.c 	er = kvmppc_emulate_instruction(run, vcpu);
run              1106 arch/powerpc/kvm/book3s_pr.c 		run->exit_reason = KVM_EXIT_MMIO;
run              1119 arch/powerpc/kvm/book3s_pr.c int kvmppc_handle_exit_pr(struct kvm_run *run, struct kvm_vcpu *vcpu,
run              1127 arch/powerpc/kvm/book3s_pr.c 	run->exit_reason = KVM_EXIT_UNKNOWN;
run              1128 arch/powerpc/kvm/book3s_pr.c 	run->ready_for_interrupt_injection = 1;
run              1165 arch/powerpc/kvm/book3s_pr.c 			r = kvmppc_handle_pagefault(run, vcpu, kvmppc_get_pc(vcpu), exit_nr);
run              1215 arch/powerpc/kvm/book3s_pr.c 			r = kvmppc_handle_pagefault(run, vcpu, dar, exit_nr);
run              1259 arch/powerpc/kvm/book3s_pr.c 		r = kvmppc_exit_pr_progint(run, vcpu, exit_nr);
run              1291 arch/powerpc/kvm/book3s_pr.c 			run->papr_hcall.nr = cmd;
run              1294 arch/powerpc/kvm/book3s_pr.c 				run->papr_hcall.args[i] = gpr;
run              1296 arch/powerpc/kvm/book3s_pr.c 			run->exit_reason = KVM_EXIT_PAPR_HCALL;
run              1303 arch/powerpc/kvm/book3s_pr.c 			u64 *gprs = run->osi.gprs;
run              1306 arch/powerpc/kvm/book3s_pr.c 			run->exit_reason = KVM_EXIT_OSI;
run              1337 arch/powerpc/kvm/book3s_pr.c 				r = kvmppc_exit_pr_progint(run, vcpu, exit_nr);
run              1393 arch/powerpc/kvm/book3s_pr.c 			run->exit_reason = KVM_EXIT_DEBUG;
run               719 arch/powerpc/kvm/booke.c 		vcpu->run->exit_reason = KVM_EXIT_WATCHDOG;
run               724 arch/powerpc/kvm/booke.c 		vcpu->run->epr.epr = 0;
run               726 arch/powerpc/kvm/booke.c 		vcpu->run->exit_reason = KVM_EXIT_EPR;
run               803 arch/powerpc/kvm/booke.c static int emulation_exit(struct kvm_run *run, struct kvm_vcpu *vcpu)
run               807 arch/powerpc/kvm/booke.c 	er = kvmppc_emulate_instruction(run, vcpu);
run               824 arch/powerpc/kvm/booke.c 		run->hw.hardware_exit_reason = ~0ULL << 32;
run               825 arch/powerpc/kvm/booke.c 		run->hw.hardware_exit_reason |= vcpu->arch.last_inst;
run               837 arch/powerpc/kvm/booke.c static int kvmppc_handle_debug(struct kvm_run *run, struct kvm_vcpu *vcpu)
run               869 arch/powerpc/kvm/booke.c 	run->debug.arch.status = 0;
run               870 arch/powerpc/kvm/booke.c 	run->debug.arch.address = vcpu->arch.regs.nip;
run               873 arch/powerpc/kvm/booke.c 		run->debug.arch.status |= KVMPPC_DEBUG_BREAKPOINT;
run               876 arch/powerpc/kvm/booke.c 			run->debug.arch.status |= KVMPPC_DEBUG_WATCH_WRITE;
run               878 arch/powerpc/kvm/booke.c 			run->debug.arch.status |= KVMPPC_DEBUG_WATCH_READ;
run               880 arch/powerpc/kvm/booke.c 			run->debug.arch.address = dbg_reg->dac1;
run               882 arch/powerpc/kvm/booke.c 			run->debug.arch.address = dbg_reg->dac2;
run               957 arch/powerpc/kvm/booke.c static int kvmppc_resume_inst_load(struct kvm_run *run, struct kvm_vcpu *vcpu,
run               969 arch/powerpc/kvm/booke.c 		run->hw.hardware_exit_reason = ~0ULL << 32;
run               970 arch/powerpc/kvm/booke.c 		run->hw.hardware_exit_reason |= last_inst;
run               984 arch/powerpc/kvm/booke.c int kvmppc_handle_exit(struct kvm_run *run, struct kvm_vcpu *vcpu,
run              1023 arch/powerpc/kvm/booke.c 	run->exit_reason = KVM_EXIT_UNKNOWN;
run              1024 arch/powerpc/kvm/booke.c 	run->ready_for_interrupt_injection = 1;
run              1027 arch/powerpc/kvm/booke.c 		r = kvmppc_resume_inst_load(run, vcpu, emulated, last_inst);
run              1036 arch/powerpc/kvm/booke.c 		run->hw.hardware_exit_reason = ~1ULL << 32;
run              1037 arch/powerpc/kvm/booke.c 		run->hw.hardware_exit_reason |= mfspr(SPRN_MCSR);
run              1087 arch/powerpc/kvm/booke.c 		r = emulation_exit(run, vcpu);
run              1097 arch/powerpc/kvm/booke.c 			r = kvmppc_handle_debug(run, vcpu);
run              1098 arch/powerpc/kvm/booke.c 			run->exit_reason = KVM_EXIT_DEBUG;
run              1118 arch/powerpc/kvm/booke.c 		r = emulation_exit(run, vcpu);
run              1165 arch/powerpc/kvm/booke.c 		run->hw.hardware_exit_reason = exit_nr;
run              1285 arch/powerpc/kvm/booke.c 			r = kvmppc_emulate_mmio(run, vcpu);
run              1336 arch/powerpc/kvm/booke.c 		r = kvmppc_handle_debug(run, vcpu);
run              1338 arch/powerpc/kvm/booke.c 			run->exit_reason = KVM_EXIT_DEBUG;
run                73 arch/powerpc/kvm/booke.h int kvmppc_booke_emulate_op(struct kvm_run *run, struct kvm_vcpu *vcpu,
run                98 arch/powerpc/kvm/booke.h extern int kvmppc_core_emulate_op_e500(struct kvm_run *run,
run               106 arch/powerpc/kvm/booke.h extern int kvmppc_core_emulate_op_e500(struct kvm_run *run,
run                42 arch/powerpc/kvm/booke_emulate.c int kvmppc_booke_emulate_op(struct kvm_run *run, struct kvm_vcpu *vcpu,
run                86 arch/powerpc/kvm/e500_emulate.c static int kvmppc_e500_emul_ehpriv(struct kvm_run *run, struct kvm_vcpu *vcpu,
run                93 arch/powerpc/kvm/e500_emulate.c 		run->exit_reason = KVM_EXIT_DEBUG;
run                94 arch/powerpc/kvm/e500_emulate.c 		run->debug.arch.address = vcpu->arch.regs.nip;
run                95 arch/powerpc/kvm/e500_emulate.c 		run->debug.arch.status = 0;
run               128 arch/powerpc/kvm/e500_emulate.c int kvmppc_core_emulate_op_e500(struct kvm_run *run, struct kvm_vcpu *vcpu,
run               185 arch/powerpc/kvm/e500_emulate.c 			emulated = kvmppc_e500_emul_ehpriv(run, vcpu, inst,
run               200 arch/powerpc/kvm/e500_emulate.c 		emulated = kvmppc_booke_emulate_op(run, vcpu, inst, advance);
run               194 arch/powerpc/kvm/emulate.c int kvmppc_emulate_instruction(struct kvm_run *run, struct kvm_vcpu *vcpu)
run               273 arch/powerpc/kvm/emulate.c 			run->exit_reason = KVM_EXIT_DEBUG;
run               274 arch/powerpc/kvm/emulate.c 			run->debug.arch.status = 0;
run               275 arch/powerpc/kvm/emulate.c 			run->debug.arch.address = kvmppc_get_pc(vcpu);
run               288 arch/powerpc/kvm/emulate.c 		emulated = vcpu->kvm->arch.kvm_ops->emulate_op(run, vcpu, inst,
run                74 arch/powerpc/kvm/emulate_loadstore.c 	struct kvm_run *run = vcpu->run;
run               107 arch/powerpc/kvm/emulate_loadstore.c 				emulated = kvmppc_handle_loads(run, vcpu,
run               110 arch/powerpc/kvm/emulate_loadstore.c 				emulated = kvmppc_handle_load(run, vcpu,
run               127 arch/powerpc/kvm/emulate_loadstore.c 				emulated = kvmppc_handle_loads(run, vcpu,
run               130 arch/powerpc/kvm/emulate_loadstore.c 				emulated = kvmppc_handle_load(run, vcpu,
run               167 arch/powerpc/kvm/emulate_loadstore.c 				emulated = kvmppc_handle_vmx_load(run,
run               172 arch/powerpc/kvm/emulate_loadstore.c 				emulated = kvmppc_handle_vmx_load(run, vcpu,
run               220 arch/powerpc/kvm/emulate_loadstore.c 			emulated = kvmppc_handle_vsx_load(run, vcpu,
run               230 arch/powerpc/kvm/emulate_loadstore.c 			emulated = kvmppc_handle_store(run, vcpu, op.val,
run               253 arch/powerpc/kvm/emulate_loadstore.c 			emulated = kvmppc_handle_store(run, vcpu,
run               293 arch/powerpc/kvm/emulate_loadstore.c 				emulated = kvmppc_handle_vmx_store(run,
run               297 arch/powerpc/kvm/emulate_loadstore.c 				emulated = kvmppc_handle_vmx_store(run,
run               341 arch/powerpc/kvm/emulate_loadstore.c 			emulated = kvmppc_handle_vsx_store(run, vcpu,
run                94 arch/powerpc/kvm/powerpc.c 			vcpu->run->exit_reason = KVM_EXIT_INTR;
run               281 arch/powerpc/kvm/powerpc.c int kvmppc_emulate_mmio(struct kvm_run *run, struct kvm_vcpu *vcpu)
run               297 arch/powerpc/kvm/powerpc.c 		run->exit_reason = KVM_EXIT_MMIO;
run              1111 arch/powerpc/kvm/powerpc.c                                       struct kvm_run *run)
run              1115 arch/powerpc/kvm/powerpc.c 	if (run->mmio.len > sizeof(gpr)) {
run              1116 arch/powerpc/kvm/powerpc.c 		printk(KERN_ERR "bad MMIO length: %d\n", run->mmio.len);
run              1121 arch/powerpc/kvm/powerpc.c 		switch (run->mmio.len) {
run              1122 arch/powerpc/kvm/powerpc.c 		case 8: gpr = *(u64 *)run->mmio.data; break;
run              1123 arch/powerpc/kvm/powerpc.c 		case 4: gpr = *(u32 *)run->mmio.data; break;
run              1124 arch/powerpc/kvm/powerpc.c 		case 2: gpr = *(u16 *)run->mmio.data; break;
run              1125 arch/powerpc/kvm/powerpc.c 		case 1: gpr = *(u8 *)run->mmio.data; break;
run              1128 arch/powerpc/kvm/powerpc.c 		switch (run->mmio.len) {
run              1129 arch/powerpc/kvm/powerpc.c 		case 8: gpr = swab64(*(u64 *)run->mmio.data); break;
run              1130 arch/powerpc/kvm/powerpc.c 		case 4: gpr = swab32(*(u32 *)run->mmio.data); break;
run              1131 arch/powerpc/kvm/powerpc.c 		case 2: gpr = swab16(*(u16 *)run->mmio.data); break;
run              1132 arch/powerpc/kvm/powerpc.c 		case 1: gpr = *(u8 *)run->mmio.data; break;
run              1137 arch/powerpc/kvm/powerpc.c 	if ((vcpu->arch.mmio_sp64_extend) && (run->mmio.len == 4))
run              1141 arch/powerpc/kvm/powerpc.c 		switch (run->mmio.len) {
run              1222 arch/powerpc/kvm/powerpc.c static int __kvmppc_handle_load(struct kvm_run *run, struct kvm_vcpu *vcpu,
run              1236 arch/powerpc/kvm/powerpc.c 	if (bytes > sizeof(run->mmio.data)) {
run              1238 arch/powerpc/kvm/powerpc.c 		       run->mmio.len);
run              1241 arch/powerpc/kvm/powerpc.c 	run->mmio.phys_addr = vcpu->arch.paddr_accessed;
run              1242 arch/powerpc/kvm/powerpc.c 	run->mmio.len = bytes;
run              1243 arch/powerpc/kvm/powerpc.c 	run->mmio.is_write = 0;
run              1253 arch/powerpc/kvm/powerpc.c 	ret = kvm_io_bus_read(vcpu, KVM_MMIO_BUS, run->mmio.phys_addr,
run              1254 arch/powerpc/kvm/powerpc.c 			      bytes, &run->mmio.data);
run              1259 arch/powerpc/kvm/powerpc.c 		kvmppc_complete_mmio_load(vcpu, run);
run              1267 arch/powerpc/kvm/powerpc.c int kvmppc_handle_load(struct kvm_run *run, struct kvm_vcpu *vcpu,
run              1271 arch/powerpc/kvm/powerpc.c 	return __kvmppc_handle_load(run, vcpu, rt, bytes, is_default_endian, 0);
run              1276 arch/powerpc/kvm/powerpc.c int kvmppc_handle_loads(struct kvm_run *run, struct kvm_vcpu *vcpu,
run              1280 arch/powerpc/kvm/powerpc.c 	return __kvmppc_handle_load(run, vcpu, rt, bytes, is_default_endian, 1);
run              1284 arch/powerpc/kvm/powerpc.c int kvmppc_handle_vsx_load(struct kvm_run *run, struct kvm_vcpu *vcpu,
run              1295 arch/powerpc/kvm/powerpc.c 		emulated = __kvmppc_handle_load(run, vcpu, rt, bytes,
run              1301 arch/powerpc/kvm/powerpc.c 		vcpu->arch.paddr_accessed += run->mmio.len;
run              1310 arch/powerpc/kvm/powerpc.c int kvmppc_handle_store(struct kvm_run *run, struct kvm_vcpu *vcpu,
run              1313 arch/powerpc/kvm/powerpc.c 	void *data = run->mmio.data;
run              1324 arch/powerpc/kvm/powerpc.c 	if (bytes > sizeof(run->mmio.data)) {
run              1326 arch/powerpc/kvm/powerpc.c 		       run->mmio.len);
run              1329 arch/powerpc/kvm/powerpc.c 	run->mmio.phys_addr = vcpu->arch.paddr_accessed;
run              1330 arch/powerpc/kvm/powerpc.c 	run->mmio.len = bytes;
run              1331 arch/powerpc/kvm/powerpc.c 	run->mmio.is_write = 1;
run              1357 arch/powerpc/kvm/powerpc.c 	ret = kvm_io_bus_write(vcpu, KVM_MMIO_BUS, run->mmio.phys_addr,
run              1358 arch/powerpc/kvm/powerpc.c 			       bytes, &run->mmio.data);
run              1426 arch/powerpc/kvm/powerpc.c int kvmppc_handle_vsx_store(struct kvm_run *run, struct kvm_vcpu *vcpu,
run              1442 arch/powerpc/kvm/powerpc.c 		emulated = kvmppc_handle_store(run, vcpu,
run              1448 arch/powerpc/kvm/powerpc.c 		vcpu->arch.paddr_accessed += run->mmio.len;
run              1458 arch/powerpc/kvm/powerpc.c 			struct kvm_run *run)
run              1463 arch/powerpc/kvm/powerpc.c 	vcpu->arch.paddr_accessed += run->mmio.len;
run              1466 arch/powerpc/kvm/powerpc.c 		emulated = kvmppc_handle_vsx_load(run, vcpu, vcpu->arch.io_gpr,
run              1467 arch/powerpc/kvm/powerpc.c 			 run->mmio.len, 1, vcpu->arch.mmio_sign_extend);
run              1469 arch/powerpc/kvm/powerpc.c 		emulated = kvmppc_handle_vsx_store(run, vcpu,
run              1470 arch/powerpc/kvm/powerpc.c 			 vcpu->arch.io_gpr, run->mmio.len, 1);
run              1475 arch/powerpc/kvm/powerpc.c 		run->exit_reason = KVM_EXIT_MMIO;
run              1480 arch/powerpc/kvm/powerpc.c 		run->exit_reason = KVM_EXIT_INTERNAL_ERROR;
run              1481 arch/powerpc/kvm/powerpc.c 		run->internal.suberror = KVM_INTERNAL_ERROR_EMULATION;
run              1493 arch/powerpc/kvm/powerpc.c int kvmppc_handle_vmx_load(struct kvm_run *run, struct kvm_vcpu *vcpu,
run              1502 arch/powerpc/kvm/powerpc.c 		emulated = __kvmppc_handle_load(run, vcpu, rt, bytes,
run              1508 arch/powerpc/kvm/powerpc.c 		vcpu->arch.paddr_accessed += run->mmio.len;
run              1588 arch/powerpc/kvm/powerpc.c int kvmppc_handle_vmx_store(struct kvm_run *run, struct kvm_vcpu *vcpu,
run              1623 arch/powerpc/kvm/powerpc.c 		emulated = kvmppc_handle_store(run, vcpu, val, bytes,
run              1628 arch/powerpc/kvm/powerpc.c 		vcpu->arch.paddr_accessed += run->mmio.len;
run              1637 arch/powerpc/kvm/powerpc.c 		struct kvm_run *run)
run              1642 arch/powerpc/kvm/powerpc.c 	vcpu->arch.paddr_accessed += run->mmio.len;
run              1645 arch/powerpc/kvm/powerpc.c 		emulated = kvmppc_handle_vmx_load(run, vcpu,
run              1646 arch/powerpc/kvm/powerpc.c 				vcpu->arch.io_gpr, run->mmio.len, 1);
run              1648 arch/powerpc/kvm/powerpc.c 		emulated = kvmppc_handle_vmx_store(run, vcpu,
run              1649 arch/powerpc/kvm/powerpc.c 				vcpu->arch.io_gpr, run->mmio.len, 1);
run              1654 arch/powerpc/kvm/powerpc.c 		run->exit_reason = KVM_EXIT_MMIO;
run              1659 arch/powerpc/kvm/powerpc.c 		run->exit_reason = KVM_EXIT_INTERNAL_ERROR;
run              1660 arch/powerpc/kvm/powerpc.c 		run->internal.suberror = KVM_INTERNAL_ERROR_EMULATION;
run              1768 arch/powerpc/kvm/powerpc.c int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu, struct kvm_run *run)
run              1777 arch/powerpc/kvm/powerpc.c 			kvmppc_complete_mmio_load(vcpu, run);
run              1785 arch/powerpc/kvm/powerpc.c 			r = kvmppc_emulate_mmio_vsx_loadstore(vcpu, run);
run              1799 arch/powerpc/kvm/powerpc.c 			r = kvmppc_emulate_mmio_vmx_loadstore(vcpu, run);
run              1807 arch/powerpc/kvm/powerpc.c 		u64 *gprs = run->osi.gprs;
run              1816 arch/powerpc/kvm/powerpc.c 		kvmppc_set_gpr(vcpu, 3, run->papr_hcall.ret);
run              1818 arch/powerpc/kvm/powerpc.c 			kvmppc_set_gpr(vcpu, 4 + i, run->papr_hcall.args[i]);
run              1822 arch/powerpc/kvm/powerpc.c 		kvmppc_set_epr(vcpu, run->epr.epr);
run              1829 arch/powerpc/kvm/powerpc.c 	if (run->immediate_exit)
run              1832 arch/powerpc/kvm/powerpc.c 		r = kvmppc_vcpu_run(run, vcpu);
run               475 arch/powerpc/kvm/trace_hv.h 	TP_PROTO(struct kvm_vcpu *vcpu, struct kvm_run *run),
run               477 arch/powerpc/kvm/trace_hv.h 	TP_ARGS(vcpu, run),
run               487 arch/powerpc/kvm/trace_hv.h 		__entry->exit     = run->exit_reason;
run                26 arch/s390/kvm/diag.c 	start = vcpu->run->s.regs.gprs[(vcpu->arch.sie_block->ipa & 0xf0) >> 4];
run                27 arch/s390/kvm/diag.c 	end = vcpu->run->s.regs.gprs[vcpu->arch.sie_block->ipa & 0xf] + PAGE_SIZE;
run                77 arch/s390/kvm/diag.c 		   vcpu->run->s.regs.gprs[rx]);
run                79 arch/s390/kvm/diag.c 	if (vcpu->run->s.regs.gprs[rx] & 7)
run                81 arch/s390/kvm/diag.c 	rc = read_guest(vcpu, vcpu->run->s.regs.gprs[rx], rx, &parm, sizeof(parm));
run                98 arch/s390/kvm/diag.c 			vcpu->run->s.regs.gprs[ry] = 8;
run               112 arch/s390/kvm/diag.c 		vcpu->run->s.regs.gprs[ry] = 0;
run               126 arch/s390/kvm/diag.c 		vcpu->run->s.regs.gprs[ry] = 0;
run               132 arch/s390/kvm/diag.c 			vcpu->run->s.regs.gprs[ry] = 4;
run               159 arch/s390/kvm/diag.c 	tid = vcpu->run->s.regs.gprs[(vcpu->arch.sie_block->ipa & 0xf0) >> 4];
run               175 arch/s390/kvm/diag.c 	unsigned long subcode = vcpu->run->s.regs.gprs[reg] & 0xffff;
run               181 arch/s390/kvm/diag.c 		vcpu->run->s390_reset_flags = KVM_S390_RESET_CLEAR;
run               184 arch/s390/kvm/diag.c 		vcpu->run->s390_reset_flags = 0;
run               192 arch/s390/kvm/diag.c 	vcpu->run->s390_reset_flags |= KVM_S390_RESET_SUBSYSTEM;
run               193 arch/s390/kvm/diag.c 	vcpu->run->s390_reset_flags |= KVM_S390_RESET_IPL;
run               194 arch/s390/kvm/diag.c 	vcpu->run->s390_reset_flags |= KVM_S390_RESET_CPU_INIT;
run               195 arch/s390/kvm/diag.c 	vcpu->run->exit_reason = KVM_EXIT_S390_RESET;
run               197 arch/s390/kvm/diag.c 	  vcpu->run->s390_reset_flags);
run               198 arch/s390/kvm/diag.c 	trace_kvm_s390_request_resets(vcpu->run->s390_reset_flags);
run               209 arch/s390/kvm/diag.c 	    (vcpu->run->s.regs.gprs[1] != KVM_S390_VIRTIO_CCW_NOTIFY))
run               213 arch/s390/kvm/diag.c 			    (u32) vcpu->run->s.regs.gprs[2],
run               214 arch/s390/kvm/diag.c 			    (u32) vcpu->run->s.regs.gprs[3],
run               215 arch/s390/kvm/diag.c 			    vcpu->run->s.regs.gprs[4]);
run               224 arch/s390/kvm/diag.c 				      vcpu->run->s.regs.gprs[2] & 0xffffffff,
run               225 arch/s390/kvm/diag.c 				      8, &vcpu->run->s.regs.gprs[3],
run               226 arch/s390/kvm/diag.c 				      vcpu->run->s.regs.gprs[4]);
run               233 arch/s390/kvm/diag.c 		vcpu->run->s.regs.gprs[2] = ret;
run               393 arch/s390/kvm/gaccess.c 	save_access_regs(vcpu->run->s.regs.acrs);
run               394 arch/s390/kvm/gaccess.c 	alet.val = vcpu->run->s.regs.acrs[ar];
run               373 arch/s390/kvm/guestdbg.c 	vcpu->run->exit_reason = KVM_EXIT_DEBUG;
run               391 arch/s390/kvm/guestdbg.c 	struct kvm_debug_exit_arch *debug_exit = &vcpu->run->debug.arch;
run               485 arch/s390/kvm/guestdbg.c 			*addr = base ? vcpu->run->s.regs.gprs[base] : 0;
run               486 arch/s390/kvm/guestdbg.c 			*addr += index ? vcpu->run->s.regs.gprs[index] : 0;
run               322 arch/s390/kvm/intercept.c 	rc = guest_translate_address(vcpu, vcpu->run->s.regs.gprs[reg2],
run               331 arch/s390/kvm/intercept.c 	rc = guest_translate_address(vcpu, vcpu->run->s.regs.gprs[reg1],
run               371 arch/s390/kvm/intercept.c 	code = vcpu->run->s.regs.gprs[reg1];
run               372 arch/s390/kvm/intercept.c 	addr = vcpu->run->s.regs.gprs[reg2];
run               406 arch/s390/kvm/intercept.c 	vcpu->run->s.regs.gprs[reg2 + 1] = rc;
run               559 arch/s390/kvm/interrupt.c 	save_access_regs(vcpu->run->s.regs.acrs);
run               589 arch/s390/kvm/interrupt.c 		if (write_guest_abs(vcpu, ext_sa_addr, vcpu->run->s.regs.vrs,
run               598 arch/s390/kvm/interrupt.c 				    &vcpu->run->s.regs.gscb, 32))
run               614 arch/s390/kvm/interrupt.c 		convert_vx_to_fp(fprs, (__vector128 *) vcpu->run->s.regs.vrs);
run               618 arch/s390/kvm/interrupt.c 				     vcpu->run->s.regs.fprs, 128);
run               621 arch/s390/kvm/interrupt.c 			     vcpu->run->s.regs.gprs, 128);
run               631 arch/s390/kvm/interrupt.c 			     &vcpu->run->s.regs.acrs, 64);
run              2716 arch/s390/kvm/kvm-s390.c 	vcpu->run->kvm_valid_regs = KVM_SYNC_PREFIX |
run              2724 arch/s390/kvm/kvm-s390.c 		vcpu->run->kvm_valid_regs |= KVM_SYNC_RICCB;
run              2726 arch/s390/kvm/kvm-s390.c 		vcpu->run->kvm_valid_regs |= KVM_SYNC_BPBC;
run              2728 arch/s390/kvm/kvm-s390.c 		vcpu->run->kvm_valid_regs |= KVM_SYNC_GSCB;
run              2730 arch/s390/kvm/kvm-s390.c 		vcpu->run->kvm_valid_regs |= KVM_SYNC_ETOKEN;
run              2735 arch/s390/kvm/kvm-s390.c 		vcpu->run->kvm_valid_regs |= KVM_SYNC_VRS;
run              2737 arch/s390/kvm/kvm-s390.c 		vcpu->run->kvm_valid_regs |= KVM_SYNC_FPRS;
run              2869 arch/s390/kvm/kvm-s390.c 	vcpu->run->s.regs.fpc = 0;
run              3018 arch/s390/kvm/kvm-s390.c 	vcpu->arch.sie_block->sdnxo = ((unsigned long) &vcpu->run->s.regs.sdnx)
run              3020 arch/s390/kvm/kvm-s390.c 	vcpu->arch.sie_block->riccbd = (unsigned long) &vcpu->run->s.regs.riccb;
run              3303 arch/s390/kvm/kvm-s390.c 	memcpy(&vcpu->run->s.regs.gprs, &regs->gprs, sizeof(regs->gprs));
run              3311 arch/s390/kvm/kvm-s390.c 	memcpy(&regs->gprs, &vcpu->run->s.regs.gprs, sizeof(regs->gprs));
run              3321 arch/s390/kvm/kvm-s390.c 	memcpy(&vcpu->run->s.regs.acrs, &sregs->acrs, sizeof(sregs->acrs));
run              3333 arch/s390/kvm/kvm-s390.c 	memcpy(&sregs->acrs, &vcpu->run->s.regs.acrs, sizeof(sregs->acrs));
run              3350 arch/s390/kvm/kvm-s390.c 	vcpu->run->s.regs.fpc = fpu->fpc;
run              3352 arch/s390/kvm/kvm-s390.c 		convert_fp_to_vx((__vector128 *) vcpu->run->s.regs.vrs,
run              3355 arch/s390/kvm/kvm-s390.c 		memcpy(vcpu->run->s.regs.fprs, &fpu->fprs, sizeof(fpu->fprs));
run              3370 arch/s390/kvm/kvm-s390.c 				 (__vector128 *) vcpu->run->s.regs.vrs);
run              3372 arch/s390/kvm/kvm-s390.c 		memcpy(fpu->fprs, vcpu->run->s.regs.fprs, sizeof(fpu->fprs));
run              3373 arch/s390/kvm/kvm-s390.c 	fpu->fpc = vcpu->run->s.regs.fpc;
run              3386 arch/s390/kvm/kvm-s390.c 		vcpu->run->psw_mask = psw.mask;
run              3387 arch/s390/kvm/kvm-s390.c 		vcpu->run->psw_addr = psw.addr;
run              3705 arch/s390/kvm/kvm-s390.c 	vcpu->arch.sie_block->gg14 = vcpu->run->s.regs.gprs[14];
run              3706 arch/s390/kvm/kvm-s390.c 	vcpu->arch.sie_block->gg15 = vcpu->run->s.regs.gprs[15];
run              3787 arch/s390/kvm/kvm-s390.c 	vcpu->run->s.regs.gprs[14] = vcpu->arch.sie_block->gg14;
run              3788 arch/s390/kvm/kvm-s390.c 	vcpu->run->s.regs.gprs[15] = vcpu->arch.sie_block->gg15;
run              3804 arch/s390/kvm/kvm-s390.c 		vcpu->run->exit_reason = KVM_EXIT_S390_SIEIC;
run              3805 arch/s390/kvm/kvm-s390.c 		vcpu->run->s390_sieic.icptcode = vcpu->arch.sie_block->icptcode;
run              3806 arch/s390/kvm/kvm-s390.c 		vcpu->run->s390_sieic.ipa = vcpu->arch.sie_block->ipa;
run              3807 arch/s390/kvm/kvm-s390.c 		vcpu->run->s390_sieic.ipb = vcpu->arch.sie_block->ipb;
run              3813 arch/s390/kvm/kvm-s390.c 		vcpu->run->exit_reason = KVM_EXIT_S390_UCONTROL;
run              3814 arch/s390/kvm/kvm-s390.c 		vcpu->run->s390_ucontrol.trans_exc_code =
run              3816 arch/s390/kvm/kvm-s390.c 		vcpu->run->s390_ucontrol.pgm_code = 0x10;
run              3853 arch/s390/kvm/kvm-s390.c 				     vcpu->run->s.regs.gprs);
run              3927 arch/s390/kvm/kvm-s390.c 	restore_access_regs(vcpu->run->s.regs.acrs);
run              3933 arch/s390/kvm/kvm-s390.c 		current->thread.fpu.regs = vcpu->run->s.regs.vrs;
run              3935 arch/s390/kvm/kvm-s390.c 		current->thread.fpu.regs = vcpu->run->s.regs.fprs;
run              3936 arch/s390/kvm/kvm-s390.c 	current->thread.fpu.fpc = vcpu->run->s.regs.fpc;
run              3949 arch/s390/kvm/kvm-s390.c 						&vcpu->run->s.regs.gscb;
run              3974 arch/s390/kvm/kvm-s390.c 	save_access_regs(vcpu->run->s.regs.acrs);
run              3978 arch/s390/kvm/kvm-s390.c 	vcpu->run->s.regs.fpc = current->thread.fpu.fpc;
run              4087 arch/s390/kvm/kvm-s390.c 		convert_vx_to_fp(fprs, (__vector128 *) vcpu->run->s.regs.vrs);
run              4092 arch/s390/kvm/kvm-s390.c 				     vcpu->run->s.regs.fprs, 128);
run              4095 arch/s390/kvm/kvm-s390.c 			      vcpu->run->s.regs.gprs, 128);
run              4101 arch/s390/kvm/kvm-s390.c 			      &vcpu->run->s.regs.fpc, 4);
run              4111 arch/s390/kvm/kvm-s390.c 			      &vcpu->run->s.regs.acrs, 64);
run              4125 arch/s390/kvm/kvm-s390.c 	vcpu->run->s.regs.fpc = current->thread.fpu.fpc;
run              4126 arch/s390/kvm/kvm-s390.c 	save_access_regs(vcpu->run->s.regs.acrs);
run               107 arch/s390/kvm/kvm-s390.h 	return (base2 ? vcpu->run->s.regs.gprs[base2] : 0) + disp2;
run               119 arch/s390/kvm/kvm-s390.h 	*address1 = (base1 ? vcpu->run->s.regs.gprs[base1] : 0) + disp1;
run               120 arch/s390/kvm/kvm-s390.h 	*address2 = (base2 ? vcpu->run->s.regs.gprs[base2] : 0) + disp2;
run               148 arch/s390/kvm/kvm-s390.h 	return (base2 ? vcpu->run->s.regs.gprs[base2] : 0) + (long)(int)disp2;
run               159 arch/s390/kvm/kvm-s390.h 	return (base2 ? vcpu->run->s.regs.gprs[base2] : 0) + disp2;
run                64 arch/s390/kvm/priv.c 		current->thread.gs_cb = (struct gs_cb *)&vcpu->run->s.regs.gscb;
run               265 arch/s390/kvm/priv.c 	gaddr = vcpu->run->s.regs.gprs[reg2] & PAGE_MASK;
run               289 arch/s390/kvm/priv.c 	vcpu->run->s.regs.gprs[reg1] &= ~0xff;
run               290 arch/s390/kvm/priv.c 	vcpu->run->s.regs.gprs[reg1] |= key;
run               312 arch/s390/kvm/priv.c 	gaddr = vcpu->run->s.regs.gprs[reg2] & PAGE_MASK;
run               370 arch/s390/kvm/priv.c 	key = vcpu->run->s.regs.gprs[reg1] & 0xfe;
run               371 arch/s390/kvm/priv.c 	start = vcpu->run->s.regs.gprs[reg2] & PAGE_MASK;
run               412 arch/s390/kvm/priv.c 			vcpu->run->s.regs.gprs[reg1] &= ~0xff00UL;
run               413 arch/s390/kvm/priv.c 			vcpu->run->s.regs.gprs[reg1] |= (u64) oldkey << 8;
run               418 arch/s390/kvm/priv.c 			vcpu->run->s.regs.gprs[reg2] &= ~PAGE_MASK;
run               420 arch/s390/kvm/priv.c 			vcpu->run->s.regs.gprs[reg2] &= ~0xfffff000UL;
run               422 arch/s390/kvm/priv.c 		vcpu->run->s.regs.gprs[reg2] |= end;
run               449 arch/s390/kvm/priv.c 	addr = vcpu->run->s.regs.gprs[reg2] & PAGE_MASK;
run               464 arch/s390/kvm/priv.c 	vcpu->run->s.regs.gprs[0] = 0;
run               542 arch/s390/kvm/priv.c 	if (vcpu->run->s.regs.gprs[1])
run               544 arch/s390/kvm/priv.c 					   vcpu->run->s.regs.gprs[1]);
run               554 arch/s390/kvm/priv.c 	vcpu->run->exit_reason = KVM_EXIT_S390_TSCH;
run               555 arch/s390/kvm/priv.c 	vcpu->run->s390_tsch.dequeued = !!inti;
run               557 arch/s390/kvm/priv.c 		vcpu->run->s390_tsch.subchannel_id = inti->io.subchannel_id;
run               558 arch/s390/kvm/priv.c 		vcpu->run->s390_tsch.subchannel_nr = inti->io.subchannel_nr;
run               559 arch/s390/kvm/priv.c 		vcpu->run->s390_tsch.io_int_parm = inti->io.io_int_parm;
run               560 arch/s390/kvm/priv.c 		vcpu->run->s390_tsch.io_int_word = inti->io.io_int_word;
run               562 arch/s390/kvm/priv.c 	vcpu->run->s390_tsch.ipb = vcpu->arch.sie_block->ipb;
run               632 arch/s390/kvm/priv.c 	reg0 = vcpu->run->s.regs.gprs[0];
run               666 arch/s390/kvm/priv.c 		if (!ret && vcpu->run->s.regs.gprs[1] & 0x00ff0000)
run               676 arch/s390/kvm/priv.c 	memcpy(&vcpu->run->s.regs.gprs[1], &status, sizeof(status));
run               836 arch/s390/kvm/priv.c 	vcpu->run->exit_reason = KVM_EXIT_S390_STSI;
run               837 arch/s390/kvm/priv.c 	vcpu->run->s390_stsi.addr = addr;
run               838 arch/s390/kvm/priv.c 	vcpu->run->s390_stsi.ar = ar;
run               839 arch/s390/kvm/priv.c 	vcpu->run->s390_stsi.fc = fc;
run               840 arch/s390/kvm/priv.c 	vcpu->run->s390_stsi.sel1 = sel1;
run               841 arch/s390/kvm/priv.c 	vcpu->run->s390_stsi.sel2 = sel2;
run               846 arch/s390/kvm/priv.c 	int fc = (vcpu->run->s.regs.gprs[0] & 0xf0000000) >> 28;
run               847 arch/s390/kvm/priv.c 	int sel1 = vcpu->run->s.regs.gprs[0] & 0xff;
run               848 arch/s390/kvm/priv.c 	int sel2 = vcpu->run->s.regs.gprs[1] & 0xffff;
run               865 arch/s390/kvm/priv.c 	if (vcpu->run->s.regs.gprs[0] & 0x0fffff00
run               866 arch/s390/kvm/priv.c 	    || vcpu->run->s.regs.gprs[1] & 0xffff0000)
run               870 arch/s390/kvm/priv.c 		vcpu->run->s.regs.gprs[0] = 3 << 28;
run               911 arch/s390/kvm/priv.c 	vcpu->run->s.regs.gprs[0] = 0;
run               987 arch/s390/kvm/priv.c 	vcpu->run->s.regs.gprs[reg1] &= 0xffffffff00000000UL;
run               988 arch/s390/kvm/priv.c 	vcpu->run->s.regs.gprs[reg1] |= vcpu->arch.sie_block->gpsw.mask >> 32;
run               990 arch/s390/kvm/priv.c 		vcpu->run->s.regs.gprs[reg2] &= 0xffffffff00000000UL;
run               991 arch/s390/kvm/priv.c 		vcpu->run->s.regs.gprs[reg2] |=
run              1024 arch/s390/kvm/priv.c 	if (vcpu->run->s.regs.gprs[reg1] & PFMF_RESERVED)
run              1028 arch/s390/kvm/priv.c 	if (vcpu->run->s.regs.gprs[reg1] & PFMF_NQ &&
run              1033 arch/s390/kvm/priv.c 	if (vcpu->run->s.regs.gprs[reg1] & PFMF_SK &&
run              1035 arch/s390/kvm/priv.c 		mr = vcpu->run->s.regs.gprs[reg1] & PFMF_MR;
run              1036 arch/s390/kvm/priv.c 		mc = vcpu->run->s.regs.gprs[reg1] & PFMF_MC;
run              1039 arch/s390/kvm/priv.c 	nq = vcpu->run->s.regs.gprs[reg1] & PFMF_NQ;
run              1040 arch/s390/kvm/priv.c 	key = vcpu->run->s.regs.gprs[reg1] & PFMF_KEY;
run              1041 arch/s390/kvm/priv.c 	start = vcpu->run->s.regs.gprs[reg2] & PAGE_MASK;
run              1044 arch/s390/kvm/priv.c 	if (vcpu->run->s.regs.gprs[reg1] & PFMF_CF) {
run              1049 arch/s390/kvm/priv.c 	switch (vcpu->run->s.regs.gprs[reg1] & PFMF_FSC) {
run              1079 arch/s390/kvm/priv.c 		if (vcpu->run->s.regs.gprs[reg1] & PFMF_CF) {
run              1084 arch/s390/kvm/priv.c 		if (vcpu->run->s.regs.gprs[reg1] & PFMF_SK) {
run              1107 arch/s390/kvm/priv.c 	if (vcpu->run->s.regs.gprs[reg1] & PFMF_FSC) {
run              1109 arch/s390/kvm/priv.c 			vcpu->run->s.regs.gprs[reg2] = end;
run              1111 arch/s390/kvm/priv.c 			vcpu->run->s.regs.gprs[reg2] &= ~0xffffffffUL;
run              1113 arch/s390/kvm/priv.c 			vcpu->run->s.regs.gprs[reg2] |= end;
run              1134 arch/s390/kvm/priv.c 	gfn = vcpu->run->s.regs.gprs[r2] >> PAGE_SHIFT;
run              1144 arch/s390/kvm/priv.c 		vcpu->run->s.regs.gprs[r1] = res; /* Exception Indication */
run              1161 arch/s390/kvm/priv.c 	vcpu->run->s.regs.gprs[r1] = res;
run              1509 arch/s390/kvm/priv.c 	if (vcpu->run->s.regs.gprs[0] & 0x00000000ffff0000)
run              1513 arch/s390/kvm/priv.c 	value = vcpu->run->s.regs.gprs[0] & 0x000000000000ffff;
run               414 arch/s390/kvm/sigp.c 	u16 cpu_addr = vcpu->run->s.regs.gprs[r3];
run               427 arch/s390/kvm/sigp.c 		parameter = vcpu->run->s.regs.gprs[r1];
run               429 arch/s390/kvm/sigp.c 		parameter = vcpu->run->s.regs.gprs[r1 + 1];
run               436 arch/s390/kvm/sigp.c 				     &vcpu->run->s.regs.gprs[r1]);
run               441 arch/s390/kvm/sigp.c 				     &vcpu->run->s.regs.gprs[r1]);
run               463 arch/s390/kvm/sigp.c 	u16 cpu_addr = vcpu->run->s.regs.gprs[r3];
run              1036 arch/s390/kvm/vsie.c 		rc = sie64a(scb_s, vcpu->run->s.regs.gprs);
run              1076 arch/x86/include/asm/kvm_host.h 	void (*run)(struct kvm_vcpu *vcpu);
run              1556 arch/x86/kvm/hyperv.c 	return kvm_hv_hypercall_complete(vcpu, vcpu->run->hyperv.u.hcall.result);
run              1660 arch/x86/kvm/hyperv.c 		vcpu->run->exit_reason = KVM_EXIT_HYPERV;
run              1661 arch/x86/kvm/hyperv.c 		vcpu->run->hyperv.type = KVM_EXIT_HYPERV_HCALL;
run              1662 arch/x86/kvm/hyperv.c 		vcpu->run->hyperv.u.hcall.input = param;
run              1663 arch/x86/kvm/hyperv.c 		vcpu->run->hyperv.u.hcall.params[0] = ingpa;
run              1664 arch/x86/kvm/hyperv.c 		vcpu->run->hyperv.u.hcall.params[1] = outgpa;
run              1247 arch/x86/kvm/lapic.c 	struct kvm_run *run = vcpu->run;
run              1250 arch/x86/kvm/lapic.c 	run->tpr_access.rip = kvm_rip_read(vcpu);
run              1251 arch/x86/kvm/lapic.c 	run->tpr_access.is_write = write;
run              2772 arch/x86/kvm/svm.c 	struct kvm_run *kvm_run = svm->vcpu.run;
run              2802 arch/x86/kvm/svm.c 	struct kvm_run *kvm_run = svm->vcpu.run;
run              2910 arch/x86/kvm/svm.c 	struct kvm_run *kvm_run = svm->vcpu.run;
run              4139 arch/x86/kvm/svm.c 	struct kvm_run *kvm_run = svm->vcpu.run;
run              4979 arch/x86/kvm/svm.c 	struct kvm_run *kvm_run = vcpu->run;
run              5038 arch/x86/kvm/svm.c 		vcpu->run->exit_reason = KVM_EXIT_INTERNAL_ERROR;
run              5039 arch/x86/kvm/svm.c 		vcpu->run->internal.suberror =
run              5041 arch/x86/kvm/svm.c 		vcpu->run->internal.ndata = 1;
run              5042 arch/x86/kvm/svm.c 		vcpu->run->internal.data[0] = exit_code;
run              7294 arch/x86/kvm/svm.c 	.run = svm_vcpu_run,
run              2958 arch/x86/kvm/vmx/nested.c 			vcpu->run->exit_reason = KVM_EXIT_INTERNAL_ERROR;
run              2959 arch/x86/kvm/vmx/nested.c 			vcpu->run->internal.suberror =
run              2961 arch/x86/kvm/vmx/nested.c 			vcpu->run->internal.ndata = 0;
run              4597 arch/x86/kvm/vmx/vmx.c 	struct kvm_run *kvm_run = vcpu->run;
run              4637 arch/x86/kvm/vmx/vmx.c 		vcpu->run->exit_reason = KVM_EXIT_INTERNAL_ERROR;
run              4638 arch/x86/kvm/vmx/vmx.c 		vcpu->run->internal.suberror = KVM_INTERNAL_ERROR_SIMUL_EX;
run              4639 arch/x86/kvm/vmx/vmx.c 		vcpu->run->internal.ndata = 3;
run              4640 arch/x86/kvm/vmx/vmx.c 		vcpu->run->internal.data[0] = vect_info;
run              4641 arch/x86/kvm/vmx/vmx.c 		vcpu->run->internal.data[1] = intr_info;
run              4642 arch/x86/kvm/vmx/vmx.c 		vcpu->run->internal.data[2] = error_code;
run              4707 arch/x86/kvm/vmx/vmx.c 	vcpu->run->exit_reason = KVM_EXIT_SHUTDOWN;
run              4841 arch/x86/kvm/vmx/vmx.c 				vcpu->run->exit_reason = KVM_EXIT_SET_TPR;
run              4875 arch/x86/kvm/vmx/vmx.c 	vcpu->run->exit_reason = 0;
run              4904 arch/x86/kvm/vmx/vmx.c 			vcpu->run->debug.arch.dr6 = vcpu->arch.dr6;
run              4905 arch/x86/kvm/vmx/vmx.c 			vcpu->run->debug.arch.dr7 = dr7;
run              4906 arch/x86/kvm/vmx/vmx.c 			vcpu->run->debug.arch.pc = kvm_get_linear_rip(vcpu);
run              4907 arch/x86/kvm/vmx/vmx.c 			vcpu->run->debug.arch.exception = DB_VECTOR;
run              4908 arch/x86/kvm/vmx/vmx.c 			vcpu->run->exit_reason = KVM_EXIT_DEBUG;
run              5247 arch/x86/kvm/vmx/vmx.c 			vcpu->run->exit_reason = KVM_EXIT_INTERNAL_ERROR;
run              5248 arch/x86/kvm/vmx/vmx.c 			vcpu->run->internal.suberror =
run              5250 arch/x86/kvm/vmx/vmx.c 			vcpu->run->internal.ndata = 0;
run              5871 arch/x86/kvm/vmx/vmx.c 		vcpu->run->exit_reason = KVM_EXIT_FAIL_ENTRY;
run              5872 arch/x86/kvm/vmx/vmx.c 		vcpu->run->fail_entry.hardware_entry_failure_reason
run              5879 arch/x86/kvm/vmx/vmx.c 		vcpu->run->exit_reason = KVM_EXIT_FAIL_ENTRY;
run              5880 arch/x86/kvm/vmx/vmx.c 		vcpu->run->fail_entry.hardware_entry_failure_reason
run              5897 arch/x86/kvm/vmx/vmx.c 		vcpu->run->exit_reason = KVM_EXIT_INTERNAL_ERROR;
run              5898 arch/x86/kvm/vmx/vmx.c 		vcpu->run->internal.suberror = KVM_INTERNAL_ERROR_DELIVERY_EV;
run              5899 arch/x86/kvm/vmx/vmx.c 		vcpu->run->internal.ndata = 3;
run              5900 arch/x86/kvm/vmx/vmx.c 		vcpu->run->internal.data[0] = vectoring_info;
run              5901 arch/x86/kvm/vmx/vmx.c 		vcpu->run->internal.data[1] = exit_reason;
run              5902 arch/x86/kvm/vmx/vmx.c 		vcpu->run->internal.data[2] = vcpu->arch.exit_qualification;
run              5904 arch/x86/kvm/vmx/vmx.c 			vcpu->run->internal.ndata++;
run              5905 arch/x86/kvm/vmx/vmx.c 			vcpu->run->internal.data[3] =
run              5937 arch/x86/kvm/vmx/vmx.c 		vcpu->run->exit_reason = KVM_EXIT_INTERNAL_ERROR;
run              5938 arch/x86/kvm/vmx/vmx.c 		vcpu->run->internal.suberror =
run              5940 arch/x86/kvm/vmx/vmx.c 		vcpu->run->internal.ndata = 1;
run              5941 arch/x86/kvm/vmx/vmx.c 		vcpu->run->internal.data[0] = exit_reason;
run              7839 arch/x86/kvm/vmx/vmx.c 	.run = vmx_vcpu_run,
run              5675 arch/x86/kvm/x86.c 	memcpy(vcpu->run->mmio.data, frag->data, min(8u, frag->len));
run              5791 arch/x86/kvm/x86.c 	vcpu->run->mmio.len = min(8u, vcpu->mmio_fragments[0].len);
run              5792 arch/x86/kvm/x86.c 	vcpu->run->mmio.is_write = vcpu->mmio_is_write = ops->write;
run              5793 arch/x86/kvm/x86.c 	vcpu->run->exit_reason = KVM_EXIT_MMIO;
run              5794 arch/x86/kvm/x86.c 	vcpu->run->mmio.phys_addr = gpa;
run              5925 arch/x86/kvm/x86.c 	vcpu->run->exit_reason = KVM_EXIT_IO;
run              5926 arch/x86/kvm/x86.c 	vcpu->run->io.direction = in ? KVM_EXIT_IO_IN : KVM_EXIT_IO_OUT;
run              5927 arch/x86/kvm/x86.c 	vcpu->run->io.size = size;
run              5928 arch/x86/kvm/x86.c 	vcpu->run->io.data_offset = KVM_PIO_PAGE_OFFSET * PAGE_SIZE;
run              5929 arch/x86/kvm/x86.c 	vcpu->run->io.count = count;
run              5930 arch/x86/kvm/x86.c 	vcpu->run->io.port = port;
run              6416 arch/x86/kvm/x86.c 		vcpu->run->exit_reason = KVM_EXIT_INTERNAL_ERROR;
run              6417 arch/x86/kvm/x86.c 		vcpu->run->internal.suberror = KVM_INTERNAL_ERROR_EMULATION;
run              6418 arch/x86/kvm/x86.c 		vcpu->run->internal.ndata = 0;
run              6425 arch/x86/kvm/x86.c 		vcpu->run->exit_reason = KVM_EXIT_INTERNAL_ERROR;
run              6426 arch/x86/kvm/x86.c 		vcpu->run->internal.suberror = KVM_INTERNAL_ERROR_EMULATION;
run              6427 arch/x86/kvm/x86.c 		vcpu->run->internal.ndata = 0;
run              6588 arch/x86/kvm/x86.c 	struct kvm_run *kvm_run = vcpu->run;
run              6628 arch/x86/kvm/x86.c 		struct kvm_run *kvm_run = vcpu->run;
run              7354 arch/x86/kvm/x86.c 		vcpu->run->exit_reason = KVM_EXIT_HLT;
run              7529 arch/x86/kvm/x86.c 	return vcpu->run->request_interrupt_window &&
run              7535 arch/x86/kvm/x86.c 	struct kvm_run *kvm_run = vcpu->run;
run              8065 arch/x86/kvm/x86.c 			vcpu->run->exit_reason = KVM_EXIT_TPR_ACCESS;
run              8070 arch/x86/kvm/x86.c 			vcpu->run->exit_reason = KVM_EXIT_SHUTDOWN;
run              8095 arch/x86/kvm/x86.c 				vcpu->run->exit_reason = KVM_EXIT_IOAPIC_EOI;
run              8096 arch/x86/kvm/x86.c 				vcpu->run->eoi.vector =
run              8109 arch/x86/kvm/x86.c 			vcpu->run->exit_reason = KVM_EXIT_SYSTEM_EVENT;
run              8110 arch/x86/kvm/x86.c 			vcpu->run->system_event.type = KVM_SYSTEM_EVENT_CRASH;
run              8115 arch/x86/kvm/x86.c 			vcpu->run->exit_reason = KVM_EXIT_SYSTEM_EVENT;
run              8116 arch/x86/kvm/x86.c 			vcpu->run->system_event.type = KVM_SYSTEM_EVENT_RESET;
run              8121 arch/x86/kvm/x86.c 			vcpu->run->exit_reason = KVM_EXIT_HYPERV;
run              8122 arch/x86/kvm/x86.c 			vcpu->run->hyperv = vcpu->arch.hyperv.exit;
run              8253 arch/x86/kvm/x86.c 	kvm_x86_ops->run(vcpu);
run              8408 arch/x86/kvm/x86.c 			vcpu->run->exit_reason = KVM_EXIT_IRQ_WINDOW_OPEN;
run              8417 arch/x86/kvm/x86.c 			vcpu->run->exit_reason = KVM_EXIT_INTR;
run              8470 arch/x86/kvm/x86.c 	struct kvm_run *run = vcpu->run;
run              8480 arch/x86/kvm/x86.c 		memcpy(frag->data, run->mmio.data, len);
run              8503 arch/x86/kvm/x86.c 	run->exit_reason = KVM_EXIT_MMIO;
run              8504 arch/x86/kvm/x86.c 	run->mmio.phys_addr = frag->gpa;
run              8506 arch/x86/kvm/x86.c 		memcpy(run->mmio.data, frag->data, min(8u, frag->len));
run              8507 arch/x86/kvm/x86.c 	run->mmio.len = min(8u, frag->len);
run              8508 arch/x86/kvm/x86.c 	run->mmio.is_write = vcpu->mmio_is_write;
run              8578 arch/x86/kvm/x86.c 			vcpu->run->exit_reason = KVM_EXIT_INTR;
run              8584 arch/x86/kvm/x86.c 	if (vcpu->run->kvm_valid_regs & ~KVM_SYNC_X86_VALID_FIELDS) {
run              8589 arch/x86/kvm/x86.c 	if (vcpu->run->kvm_dirty_regs) {
run              8619 arch/x86/kvm/x86.c 	if (vcpu->run->kvm_valid_regs)
run              8828 arch/x86/kvm/x86.c 		vcpu->run->exit_reason = KVM_EXIT_INTERNAL_ERROR;
run              8829 arch/x86/kvm/x86.c 		vcpu->run->internal.suberror = KVM_INTERNAL_ERROR_EMULATION;
run              8830 arch/x86/kvm/x86.c 		vcpu->run->internal.ndata = 0;
run              9089 arch/x86/kvm/x86.c 	if (vcpu->run->kvm_valid_regs & KVM_SYNC_X86_REGS)
run              9090 arch/x86/kvm/x86.c 		__get_regs(vcpu, &vcpu->run->s.regs.regs);
run              9092 arch/x86/kvm/x86.c 	if (vcpu->run->kvm_valid_regs & KVM_SYNC_X86_SREGS)
run              9093 arch/x86/kvm/x86.c 		__get_sregs(vcpu, &vcpu->run->s.regs.sregs);
run              9095 arch/x86/kvm/x86.c 	if (vcpu->run->kvm_valid_regs & KVM_SYNC_X86_EVENTS)
run              9097 arch/x86/kvm/x86.c 				vcpu, &vcpu->run->s.regs.events);
run              9102 arch/x86/kvm/x86.c 	if (vcpu->run->kvm_dirty_regs & ~KVM_SYNC_X86_VALID_FIELDS)
run              9105 arch/x86/kvm/x86.c 	if (vcpu->run->kvm_dirty_regs & KVM_SYNC_X86_REGS) {
run              9106 arch/x86/kvm/x86.c 		__set_regs(vcpu, &vcpu->run->s.regs.regs);
run              9107 arch/x86/kvm/x86.c 		vcpu->run->kvm_dirty_regs &= ~KVM_SYNC_X86_REGS;
run              9109 arch/x86/kvm/x86.c 	if (vcpu->run->kvm_dirty_regs & KVM_SYNC_X86_SREGS) {
run              9110 arch/x86/kvm/x86.c 		if (__set_sregs(vcpu, &vcpu->run->s.regs.sregs))
run              9112 arch/x86/kvm/x86.c 		vcpu->run->kvm_dirty_regs &= ~KVM_SYNC_X86_SREGS;
run              9114 arch/x86/kvm/x86.c 	if (vcpu->run->kvm_dirty_regs & KVM_SYNC_X86_EVENTS) {
run              9116 arch/x86/kvm/x86.c 				vcpu, &vcpu->run->s.regs.events))
run              9118 arch/x86/kvm/x86.c 		vcpu->run->kvm_dirty_regs &= ~KVM_SYNC_X86_EVENTS;
run               596 block/blk-mq-debugfs.c 	seq_printf(m, "%lu\n", hctx->run);
run               605 block/blk-mq-debugfs.c 	hctx->run = 0;
run               181 block/blk-mq-sched.c 	hctx->run++;
run               395 block/blk-mq-sched.c 		goto run;
run               424 block/blk-mq-sched.c 		goto run;
run               438 block/blk-mq-sched.c run:
run               154 drivers/acpi/apei/apei-base.c 	apei_exec_ins_func_t run;
run               172 drivers/acpi/apei/apei-base.c 			    !ctx->ins_table[entry->instruction].run) {
run               178 drivers/acpi/apei/apei-base.c 			run = ctx->ins_table[entry->instruction].run;
run               179 drivers/acpi/apei/apei-base.c 			rc = run(ctx, entry);
run               213 drivers/acpi/apei/apei-base.c 		if (ins >= ctx->instructions || !ins_table[ins].run) {
run                22 drivers/acpi/apei/apei-internal.h 	apei_exec_ins_func_t run;
run               111 drivers/acpi/apei/einj.c 		.run   = apei_exec_read_register,
run               115 drivers/acpi/apei/einj.c 		.run   = apei_exec_read_register_value,
run               119 drivers/acpi/apei/einj.c 		.run   = apei_exec_write_register,
run               123 drivers/acpi/apei/einj.c 		.run   = apei_exec_write_register_value,
run               127 drivers/acpi/apei/einj.c 		.run   = apei_exec_noop,
run               292 drivers/acpi/apei/erst.c 		.run = apei_exec_read_register,
run               296 drivers/acpi/apei/erst.c 		.run = apei_exec_read_register_value,
run               300 drivers/acpi/apei/erst.c 		.run = apei_exec_write_register,
run               304 drivers/acpi/apei/erst.c 		.run = apei_exec_write_register_value,
run               308 drivers/acpi/apei/erst.c 		.run = apei_exec_noop,
run               312 drivers/acpi/apei/erst.c 		.run = erst_exec_load_var1,
run               316 drivers/acpi/apei/erst.c 		.run = erst_exec_load_var2,
run               320 drivers/acpi/apei/erst.c 		.run = erst_exec_store_var1,
run               324 drivers/acpi/apei/erst.c 		.run = erst_exec_add,
run               328 drivers/acpi/apei/erst.c 		.run = erst_exec_subtract,
run               332 drivers/acpi/apei/erst.c 		.run = erst_exec_add_value,
run               336 drivers/acpi/apei/erst.c 		.run = erst_exec_subtract_value,
run               340 drivers/acpi/apei/erst.c 		.run = erst_exec_stall,
run               344 drivers/acpi/apei/erst.c 		.run = erst_exec_stall_while_true,
run               348 drivers/acpi/apei/erst.c 		.run = erst_exec_skip_next_instruction_if_true,
run               352 drivers/acpi/apei/erst.c 		.run = erst_exec_goto,
run               356 drivers/acpi/apei/erst.c 		.run = erst_exec_set_src_address_base,
run               360 drivers/acpi/apei/erst.c 		.run = erst_exec_set_dst_address_base,
run               364 drivers/acpi/apei/erst.c 		.run = erst_exec_move_data,
run               146 drivers/dma/dmatest.c module_param_cb(run, &run_ops, &dmatest_run, S_IRUGO | S_IWUSR);
run               147 drivers/dma/dmatest.c MODULE_PARM_DESC(run, "Run the test (default: false)");
run                93 drivers/gpu/drm/tiny/gm12u320.c 		bool                     run;
run               354 drivers/gpu/drm/tiny/gm12u320.c 	while (gm12u320->fb_update.run) {
run               418 drivers/gpu/drm/tiny/gm12u320.c 				   !gm12u320->fb_update.run ||
run               465 drivers/gpu/drm/tiny/gm12u320.c 	gm12u320->fb_update.run = true;
run               474 drivers/gpu/drm/tiny/gm12u320.c 	gm12u320->fb_update.run = false;
run              1241 drivers/gpu/ipu-v3/ipu-image-convert.c 	struct ipu_image_convert_run *run;
run              1246 drivers/gpu/ipu-v3/ipu-image-convert.c 	list_for_each_entry(run, q, list) {
run              1247 drivers/gpu/ipu-v3/ipu-image-convert.c 		if (run->ctx == ctx)
run              1254 drivers/gpu/ipu-v3/ipu-image-convert.c static void convert_stop(struct ipu_image_convert_run *run)
run              1256 drivers/gpu/ipu-v3/ipu-image-convert.c 	struct ipu_image_convert_ctx *ctx = run->ctx;
run              1261 drivers/gpu/ipu-v3/ipu-image-convert.c 		__func__, chan->ic_task, ctx, run);
run              1367 drivers/gpu/ipu-v3/ipu-image-convert.c static int convert_start(struct ipu_image_convert_run *run, unsigned int tile)
run              1369 drivers/gpu/ipu-v3/ipu-image-convert.c 	struct ipu_image_convert_ctx *ctx = run->ctx;
run              1381 drivers/gpu/ipu-v3/ipu-image-convert.c 		__func__, chan->ic_task, ctx, run, tile, dst_tile);
run              1479 drivers/gpu/ipu-v3/ipu-image-convert.c static int do_run(struct ipu_image_convert_run *run)
run              1481 drivers/gpu/ipu-v3/ipu-image-convert.c 	struct ipu_image_convert_ctx *ctx = run->ctx;
run              1486 drivers/gpu/ipu-v3/ipu-image-convert.c 	ctx->in.base.phys0 = run->in_phys;
run              1487 drivers/gpu/ipu-v3/ipu-image-convert.c 	ctx->out.base.phys0 = run->out_phys;
run              1493 drivers/gpu/ipu-v3/ipu-image-convert.c 	list_del(&run->list);
run              1494 drivers/gpu/ipu-v3/ipu-image-convert.c 	chan->current_run = run;
run              1496 drivers/gpu/ipu-v3/ipu-image-convert.c 	return convert_start(run, 0);
run              1503 drivers/gpu/ipu-v3/ipu-image-convert.c 	struct ipu_image_convert_run *run, *tmp;
run              1508 drivers/gpu/ipu-v3/ipu-image-convert.c 	list_for_each_entry_safe(run, tmp, &chan->pending_q, list) {
run              1510 drivers/gpu/ipu-v3/ipu-image-convert.c 		if (run->ctx->aborting) {
run              1513 drivers/gpu/ipu-v3/ipu-image-convert.c 				__func__, chan->ic_task, run->ctx, run);
run              1517 drivers/gpu/ipu-v3/ipu-image-convert.c 		ret = do_run(run);
run              1526 drivers/gpu/ipu-v3/ipu-image-convert.c 		run->status = ret;
run              1527 drivers/gpu/ipu-v3/ipu-image-convert.c 		list_add_tail(&run->list, &chan->done_q);
run              1535 drivers/gpu/ipu-v3/ipu-image-convert.c 	struct ipu_image_convert_run *run;
run              1541 drivers/gpu/ipu-v3/ipu-image-convert.c 		run = list_entry(chan->done_q.next,
run              1545 drivers/gpu/ipu-v3/ipu-image-convert.c 		list_del(&run->list);
run              1549 drivers/gpu/ipu-v3/ipu-image-convert.c 			__func__, chan->ic_task, run->ctx, run, run->status);
run              1553 drivers/gpu/ipu-v3/ipu-image-convert.c 		run->ctx->complete(run, run->ctx->complete_context);
run              1618 drivers/gpu/ipu-v3/ipu-image-convert.c static irqreturn_t do_irq(struct ipu_image_convert_run *run)
run              1620 drivers/gpu/ipu-v3/ipu-image-convert.c 	struct ipu_image_convert_ctx *ctx = run->ctx;
run              1642 drivers/gpu/ipu-v3/ipu-image-convert.c 		convert_stop(run);
run              1643 drivers/gpu/ipu-v3/ipu-image-convert.c 		run->status = -EIO;
run              1651 drivers/gpu/ipu-v3/ipu-image-convert.c 		convert_stop(run);
run              1652 drivers/gpu/ipu-v3/ipu-image-convert.c 		run->status = 0;
run              1661 drivers/gpu/ipu-v3/ipu-image-convert.c 			convert_stop(run);
run              1662 drivers/gpu/ipu-v3/ipu-image-convert.c 			convert_start(run, ctx->next_tile);
run              1706 drivers/gpu/ipu-v3/ipu-image-convert.c 	list_add_tail(&run->list, &chan->done_q);
run              1716 drivers/gpu/ipu-v3/ipu-image-convert.c 	struct ipu_image_convert_run *run;
run              1723 drivers/gpu/ipu-v3/ipu-image-convert.c 	run = chan->current_run;
run              1724 drivers/gpu/ipu-v3/ipu-image-convert.c 	if (!run) {
run              1729 drivers/gpu/ipu-v3/ipu-image-convert.c 	ctx = run->ctx;
run              1737 drivers/gpu/ipu-v3/ipu-image-convert.c 	ret = do_irq(run);
run              1748 drivers/gpu/ipu-v3/ipu-image-convert.c 	struct ipu_image_convert_run *run;
run              1755 drivers/gpu/ipu-v3/ipu-image-convert.c 	run = chan->current_run;
run              1756 drivers/gpu/ipu-v3/ipu-image-convert.c 	if (!run) {
run              1761 drivers/gpu/ipu-v3/ipu-image-convert.c 	ctx = run->ctx;
run              1770 drivers/gpu/ipu-v3/ipu-image-convert.c 	ret = do_irq(run);
run              1783 drivers/gpu/ipu-v3/ipu-image-convert.c 	struct ipu_image_convert_run *run;
run              1788 drivers/gpu/ipu-v3/ipu-image-convert.c 	run = chan->current_run;
run              1789 drivers/gpu/ipu-v3/ipu-image-convert.c 	if (run && run->ctx == ctx) {
run              1790 drivers/gpu/ipu-v3/ipu-image-convert.c 		convert_stop(run);
run              1791 drivers/gpu/ipu-v3/ipu-image-convert.c 		run->status = -EIO;
run              1792 drivers/gpu/ipu-v3/ipu-image-convert.c 		list_add_tail(&run->list, &chan->done_q);
run              2234 drivers/gpu/ipu-v3/ipu-image-convert.c int ipu_image_convert_queue(struct ipu_image_convert_run *run)
run              2242 drivers/gpu/ipu-v3/ipu-image-convert.c 	if (!run || !run->ctx || !run->in_phys || !run->out_phys)
run              2245 drivers/gpu/ipu-v3/ipu-image-convert.c 	ctx = run->ctx;
run              2250 drivers/gpu/ipu-v3/ipu-image-convert.c 		chan->ic_task, ctx, run);
run              2252 drivers/gpu/ipu-v3/ipu-image-convert.c 	INIT_LIST_HEAD(&run->list);
run              2261 drivers/gpu/ipu-v3/ipu-image-convert.c 	list_add_tail(&run->list, &chan->pending_q);
run              2264 drivers/gpu/ipu-v3/ipu-image-convert.c 		ret = do_run(run);
run              2279 drivers/gpu/ipu-v3/ipu-image-convert.c 	struct ipu_image_convert_run *run, *active_run, *tmp;
run              2286 drivers/gpu/ipu-v3/ipu-image-convert.c 	list_for_each_entry_safe(run, tmp, &chan->pending_q, list) {
run              2287 drivers/gpu/ipu-v3/ipu-image-convert.c 		if (run->ctx != ctx)
run              2289 drivers/gpu/ipu-v3/ipu-image-convert.c 		run->status = -EIO;
run              2290 drivers/gpu/ipu-v3/ipu-image-convert.c 		list_move_tail(&run->list, &chan->done_q);
run              2380 drivers/gpu/ipu-v3/ipu-image-convert.c 	struct ipu_image_convert_run *run;
run              2388 drivers/gpu/ipu-v3/ipu-image-convert.c 	run = kzalloc(sizeof(*run), GFP_KERNEL);
run              2389 drivers/gpu/ipu-v3/ipu-image-convert.c 	if (!run) {
run              2394 drivers/gpu/ipu-v3/ipu-image-convert.c 	run->ctx = ctx;
run              2395 drivers/gpu/ipu-v3/ipu-image-convert.c 	run->in_phys = in->phys0;
run              2396 drivers/gpu/ipu-v3/ipu-image-convert.c 	run->out_phys = out->phys0;
run              2398 drivers/gpu/ipu-v3/ipu-image-convert.c 	ret = ipu_image_convert_queue(run);
run              2401 drivers/gpu/ipu-v3/ipu-image-convert.c 		kfree(run);
run              2405 drivers/gpu/ipu-v3/ipu-image-convert.c 	return run;
run              2410 drivers/gpu/ipu-v3/ipu-image-convert.c static void image_convert_sync_complete(struct ipu_image_convert_run *run,
run              2422 drivers/gpu/ipu-v3/ipu-image-convert.c 	struct ipu_image_convert_run *run;
run              2428 drivers/gpu/ipu-v3/ipu-image-convert.c 	run = ipu_image_convert(ipu, ic_task, in, out, rot_mode,
run              2430 drivers/gpu/ipu-v3/ipu-image-convert.c 	if (IS_ERR(run))
run              2431 drivers/gpu/ipu-v3/ipu-image-convert.c 		return PTR_ERR(run);
run              2436 drivers/gpu/ipu-v3/ipu-image-convert.c 	ipu_image_convert_unprepare(run->ctx);
run              2437 drivers/gpu/ipu-v3/ipu-image-convert.c 	kfree(run);
run               469 drivers/iio/light/si1145.c 		int run = 1;
run               471 drivers/iio/light/si1145.c 		while (i + run < indio_dev->masklength) {
run               472 drivers/iio/light/si1145.c 			if (!test_bit(i + run, indio_dev->active_scan_mask))
run               474 drivers/iio/light/si1145.c 			if (indio_dev->channels[i + run].address !=
run               475 drivers/iio/light/si1145.c 				indio_dev->channels[i].address + 2 * run)
run               477 drivers/iio/light/si1145.c 			run++;
run               482 drivers/iio/light/si1145.c 				sizeof(u16) * run, &buffer[j]);
run               485 drivers/iio/light/si1145.c 		j += run * sizeof(u16);
run               486 drivers/iio/light/si1145.c 		i += run - 1;
run               340 drivers/md/md-faulty.c 	.run		= faulty_run,
run               319 drivers/md/md-linear.c 	.run		= linear_run,
run               474 drivers/md/md-multipath.c 	.run		= multipath_run,
run              3976 drivers/md/md.c 	pers->run(mddev);
run              5743 drivers/md/md.c 	err = pers->run(mddev);
run              7710 drivers/md/md.c 			thread->run(thread);
run              7726 drivers/md/md.c struct md_thread *md_register_thread(void (*run) (struct md_thread *),
run              7737 drivers/md/md.c 	thread->run = run;
run               558 drivers/md/md.h 	int (*run)(struct mddev *mddev);
run               666 drivers/md/md.h 	void			(*run) (struct md_thread *thread);
run               690 drivers/md/md.h 	void (*run)(struct md_thread *thread),
run               815 drivers/md/raid0.c 	.run		= raid0_run,
run              3373 drivers/md/raid1.c 	.run		= raid1_run,
run              4916 drivers/md/raid10.c 	.run		= raid10_run,
run              8415 drivers/md/raid5.c 	.run		= raid5_run,
run              8440 drivers/md/raid5.c 	.run		= raid5_run,
run              8466 drivers/md/raid5.c 	.run		= raid5_run,
run               203 drivers/media/platform/mtk-vpu/mtk_vpu.c 	struct vpu_run run;
run               419 drivers/media/platform/mtk-vpu/mtk_vpu.c 	return vpu->run.dec_capability;
run               427 drivers/media/platform/mtk-vpu/mtk_vpu.c 	return vpu->run.enc_capability;
run               533 drivers/media/platform/mtk-vpu/mtk_vpu.c 	struct vpu_run *run;
run               542 drivers/media/platform/mtk-vpu/mtk_vpu.c 	run = &vpu->run;
run               559 drivers/media/platform/mtk-vpu/mtk_vpu.c 	run->signaled = false;
run               579 drivers/media/platform/mtk-vpu/mtk_vpu.c 	ret = wait_event_interruptible_timeout(run->wq,
run               580 drivers/media/platform/mtk-vpu/mtk_vpu.c 					       run->signaled,
run               593 drivers/media/platform/mtk-vpu/mtk_vpu.c 	dev_info(dev, "vpu is ready. Fw version %s\n", run->fw_ver);
run               606 drivers/media/platform/mtk-vpu/mtk_vpu.c 	struct vpu_run *run = (struct vpu_run *)data;
run               608 drivers/media/platform/mtk-vpu/mtk_vpu.c 	vpu->run.signaled = run->signaled;
run               609 drivers/media/platform/mtk-vpu/mtk_vpu.c 	strscpy(vpu->run.fw_ver, run->fw_ver, sizeof(vpu->run.fw_ver));
run               610 drivers/media/platform/mtk-vpu/mtk_vpu.c 	vpu->run.dec_capability = run->dec_capability;
run               611 drivers/media/platform/mtk-vpu/mtk_vpu.c 	vpu->run.enc_capability = run->enc_capability;
run               612 drivers/media/platform/mtk-vpu/mtk_vpu.c 	wake_up_interruptible(&vpu->run.wq);
run               647 drivers/media/platform/mtk-vpu/mtk_vpu.c 		vpu->run.fw_ver, pc, wdt,
run               872 drivers/media/platform/mtk-vpu/mtk_vpu.c 	init_waitqueue_head(&vpu->run.wq);
run              1020 drivers/mmc/core/mmc_test.c 	int (*run)(struct mmc_test_card *);
run              2551 drivers/mmc/core/mmc_test.c 		.run = mmc_test_basic_write,
run              2556 drivers/mmc/core/mmc_test.c 		.run = mmc_test_basic_read,
run              2562 drivers/mmc/core/mmc_test.c 		.run = mmc_test_verify_write,
run              2569 drivers/mmc/core/mmc_test.c 		.run = mmc_test_verify_read,
run              2576 drivers/mmc/core/mmc_test.c 		.run = mmc_test_multi_write,
run              2583 drivers/mmc/core/mmc_test.c 		.run = mmc_test_multi_read,
run              2590 drivers/mmc/core/mmc_test.c 		.run = mmc_test_pow2_write,
run              2597 drivers/mmc/core/mmc_test.c 		.run = mmc_test_pow2_read,
run              2604 drivers/mmc/core/mmc_test.c 		.run = mmc_test_weird_write,
run              2611 drivers/mmc/core/mmc_test.c 		.run = mmc_test_weird_read,
run              2618 drivers/mmc/core/mmc_test.c 		.run = mmc_test_align_write,
run              2625 drivers/mmc/core/mmc_test.c 		.run = mmc_test_align_read,
run              2632 drivers/mmc/core/mmc_test.c 		.run = mmc_test_align_multi_write,
run              2639 drivers/mmc/core/mmc_test.c 		.run = mmc_test_align_multi_read,
run              2645 drivers/mmc/core/mmc_test.c 		.run = mmc_test_xfersize_write,
run              2650 drivers/mmc/core/mmc_test.c 		.run = mmc_test_xfersize_read,
run              2655 drivers/mmc/core/mmc_test.c 		.run = mmc_test_multi_xfersize_write,
run              2660 drivers/mmc/core/mmc_test.c 		.run = mmc_test_multi_xfersize_read,
run              2668 drivers/mmc/core/mmc_test.c 		.run = mmc_test_write_high,
run              2675 drivers/mmc/core/mmc_test.c 		.run = mmc_test_read_high,
run              2682 drivers/mmc/core/mmc_test.c 		.run = mmc_test_multi_write_high,
run              2689 drivers/mmc/core/mmc_test.c 		.run = mmc_test_multi_read_high,
run              2697 drivers/mmc/core/mmc_test.c 		.run = mmc_test_no_highmem,
run              2702 drivers/mmc/core/mmc_test.c 		.run = mmc_test_no_highmem,
run              2707 drivers/mmc/core/mmc_test.c 		.run = mmc_test_no_highmem,
run              2712 drivers/mmc/core/mmc_test.c 		.run = mmc_test_no_highmem,
run              2720 drivers/mmc/core/mmc_test.c 		.run = mmc_test_best_read_performance,
run              2727 drivers/mmc/core/mmc_test.c 		.run = mmc_test_best_write_performance,
run              2734 drivers/mmc/core/mmc_test.c 		.run = mmc_test_best_read_perf_max_scatter,
run              2741 drivers/mmc/core/mmc_test.c 		.run = mmc_test_best_write_perf_max_scatter,
run              2748 drivers/mmc/core/mmc_test.c 		.run = mmc_test_profile_read_perf,
run              2755 drivers/mmc/core/mmc_test.c 		.run = mmc_test_profile_write_perf,
run              2762 drivers/mmc/core/mmc_test.c 		.run = mmc_test_profile_trim_perf,
run              2769 drivers/mmc/core/mmc_test.c 		.run = mmc_test_profile_seq_read_perf,
run              2776 drivers/mmc/core/mmc_test.c 		.run = mmc_test_profile_seq_write_perf,
run              2783 drivers/mmc/core/mmc_test.c 		.run = mmc_test_profile_seq_trim_perf,
run              2790 drivers/mmc/core/mmc_test.c 		.run = mmc_test_random_read_perf,
run              2797 drivers/mmc/core/mmc_test.c 		.run = mmc_test_random_write_perf,
run              2804 drivers/mmc/core/mmc_test.c 		.run = mmc_test_large_seq_read_perf,
run              2811 drivers/mmc/core/mmc_test.c 		.run = mmc_test_large_seq_write_perf,
run              2818 drivers/mmc/core/mmc_test.c 		.run = mmc_test_profile_mult_write_blocking_perf,
run              2825 drivers/mmc/core/mmc_test.c 		.run = mmc_test_profile_mult_write_nonblock_perf,
run              2832 drivers/mmc/core/mmc_test.c 		.run = mmc_test_profile_mult_read_blocking_perf,
run              2839 drivers/mmc/core/mmc_test.c 		.run = mmc_test_profile_mult_read_nonblock_perf,
run              2846 drivers/mmc/core/mmc_test.c 		.run = mmc_test_profile_sglen_wr_blocking_perf,
run              2853 drivers/mmc/core/mmc_test.c 		.run = mmc_test_profile_sglen_wr_nonblock_perf,
run              2860 drivers/mmc/core/mmc_test.c 		.run = mmc_test_profile_sglen_r_blocking_perf,
run              2867 drivers/mmc/core/mmc_test.c 		.run = mmc_test_profile_sglen_r_nonblock_perf,
run              2873 drivers/mmc/core/mmc_test.c 		.run = mmc_test_reset,
run              2879 drivers/mmc/core/mmc_test.c 		.run = mmc_test_cmds_during_read,
run              2886 drivers/mmc/core/mmc_test.c 		.run = mmc_test_cmds_during_write,
run              2893 drivers/mmc/core/mmc_test.c 		.run = mmc_test_cmds_during_read_cmd23,
run              2900 drivers/mmc/core/mmc_test.c 		.run = mmc_test_cmds_during_write_cmd23,
run              2907 drivers/mmc/core/mmc_test.c 		.run = mmc_test_cmds_during_read_cmd23_nonblock,
run              2914 drivers/mmc/core/mmc_test.c 		.run = mmc_test_cmds_during_write_cmd23_nonblock,
run              2970 drivers/mmc/core/mmc_test.c 		ret = mmc_test_cases[i].run(test);
run               359 drivers/net/phy/phy_device.c 		       int (*run)(struct phy_device *))
run               369 drivers/net/phy/phy_device.c 	fixup->run = run;
run               381 drivers/net/phy/phy_device.c 			       int (*run)(struct phy_device *))
run               383 drivers/net/phy/phy_device.c 	return phy_register_fixup(PHY_ANY_ID, phy_uid, phy_uid_mask, run);
run               389 drivers/net/phy/phy_device.c 			      int (*run)(struct phy_device *))
run               391 drivers/net/phy/phy_device.c 	return phy_register_fixup(bus_id, PHY_ANY_UID, 0xffffffff, run);
run               467 drivers/net/phy/phy_device.c 			int err = fixup->run(phydev);
run               428 drivers/net/wan/farsync.c 	int run;		/* Port is running */
run               987 drivers/net/wan/farsync.c 	if (port->run)
run               997 drivers/net/wan/farsync.c 	if (port->run)
run              1336 drivers/net/wan/farsync.c 		if (!port->run)
run              1435 drivers/net/wan/farsync.c 		if (!port->run)
run              1562 drivers/net/wan/farsync.c 			if (port->run)
run              1570 drivers/net/wan/farsync.c 			if (port->run)
run              2138 drivers/net/wan/farsync.c 		if (port->run) {
run              2142 drivers/net/wan/farsync.c 			port->run = 0;
run              2150 drivers/net/wan/farsync.c 		port->run = 1;
run              2169 drivers/net/wan/farsync.c 		if (port->run) {
run              2170 drivers/net/wan/farsync.c 			port->run = 0;
run              2500 drivers/net/wan/farsync.c                 card->ports[i].run    = 0;
run              1402 drivers/net/wireless/ath/carl9170/tx.c 	bool run = false;
run              1450 drivers/net/wireless/ath/carl9170/tx.c 			run = true;
run              1457 drivers/net/wireless/ath/carl9170/tx.c 	return run;
run              1478 drivers/net/wireless/ath/carl9170/tx.c 	bool run;
run              1508 drivers/net/wireless/ath/carl9170/tx.c 		run = carl9170_tx_ampdu_queue(ar, sta, skb, info);
run              1509 drivers/net/wireless/ath/carl9170/tx.c 		if (run)
run              1078 drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c 	err = escan->run(cfg, ifp, request);
run              1126 drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c 	cfg->escan_info.run = brcmf_run_escan;
run              3300 drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c 	cfg->escan_info.run = brcmf_run_escan;
run               217 drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.h 	s32 (*run)(struct brcmf_cfg80211_info *cfg, struct brcmf_if *ifp,
run               896 drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c 		cfg->escan_info.run = brcmf_p2p_run_escan;
run               288 drivers/s390/crypto/vfio_ap_ops.c 	apqn = vcpu->run->s.regs.gprs[0] & 0xffff;
run               300 drivers/s390/crypto/vfio_ap_ops.c 	status = vcpu->run->s.regs.gprs[1];
run               305 drivers/s390/crypto/vfio_ap_ops.c 					     vcpu->run->s.regs.gprs[2]);
run               310 drivers/s390/crypto/vfio_ap_ops.c 	memcpy(&vcpu->run->s.regs.gprs[1], &qstatus, sizeof(qstatus));
run               311 drivers/s390/crypto/vfio_ap_ops.c 	vcpu->run->s.regs.gprs[1] >>= 32;
run               357 drivers/s390/scsi/zfcp_dbf.c 	rec->u.run.fsf_req_id = erp->fsf_req_id;
run               358 drivers/s390/scsi/zfcp_dbf.c 	rec->u.run.rec_status = erp->status;
run               359 drivers/s390/scsi/zfcp_dbf.c 	rec->u.run.rec_step = erp->step;
run               360 drivers/s390/scsi/zfcp_dbf.c 	rec->u.run.rec_action = erp->type;
run               363 drivers/s390/scsi/zfcp_dbf.c 		rec->u.run.rec_count =
run               366 drivers/s390/scsi/zfcp_dbf.c 		rec->u.run.rec_count = atomic_read(&erp->port->erp_counter);
run               368 drivers/s390/scsi/zfcp_dbf.c 		rec->u.run.rec_count = atomic_read(&erp->adapter->erp_counter);
run               410 drivers/s390/scsi/zfcp_dbf.c 	rec->u.run.fsf_req_id = req_id;
run               411 drivers/s390/scsi/zfcp_dbf.c 	rec->u.run.rec_status = ~0;
run               412 drivers/s390/scsi/zfcp_dbf.c 	rec->u.run.rec_step = ~0;
run               413 drivers/s390/scsi/zfcp_dbf.c 	rec->u.run.rec_action = ~0;
run               414 drivers/s390/scsi/zfcp_dbf.c 	rec->u.run.rec_count = ~0;
run                91 drivers/s390/scsi/zfcp_dbf.h 		struct zfcp_dbf_rec_running run;
run               598 drivers/scsi/aha152x.c 	void		(*run)(struct Scsi_Host *);
run              1346 drivers/scsi/aha152x.c 		INIT_WORK(&aha152x_tq, run);
run              2361 drivers/scsi/aha152x.c 		if(states[STATE].run)
run              2362 drivers/scsi/aha152x.c 			states[STATE].run(shpnt);
run              1222 drivers/staging/comedi/drivers/amplc_pci230.c 		bool run;
run              1225 drivers/staging/comedi/drivers/amplc_pci230.c 		run = pci230_handle_ao_fifo(dev, s);
run              1227 drivers/staging/comedi/drivers/amplc_pci230.c 		if (!run) {
run               334 drivers/staging/comedi/drivers/s626.h 						 * Start ADC: run A1
run               194 drivers/staging/media/hantro/hantro_drv.c 	ctx->codec_ops->run(ctx);
run               128 drivers/staging/media/hantro/hantro_hw.h 	void (*run)(struct hantro_ctx *ctx);
run               173 drivers/staging/media/hantro/rk3288_vpu_hw.c 		.run = hantro_h1_jpeg_enc_run,
run               179 drivers/staging/media/hantro/rk3288_vpu_hw.c 		.run = hantro_g1_h264_dec_run,
run               185 drivers/staging/media/hantro/rk3288_vpu_hw.c 		.run = hantro_g1_mpeg2_dec_run,
run               191 drivers/staging/media/hantro/rk3288_vpu_hw.c 		.run = hantro_g1_vp8_dec_run,
run               159 drivers/staging/media/hantro/rk3399_vpu_hw.c 		.run = rk3399_vpu_jpeg_enc_run,
run               165 drivers/staging/media/hantro/rk3399_vpu_hw.c 		.run = rk3399_vpu_mpeg2_dec_run,
run               171 drivers/staging/media/hantro/rk3399_vpu_hw.c 		.run = rk3399_vpu_vp8_dec_run,
run                88 drivers/staging/media/imx/imx-media-csc-scaler.c static void ipu_ic_pp_complete(struct ipu_image_convert_run *run, void *_ctx)
run               102 drivers/staging/media/imx/imx-media-csc-scaler.c 	v4l2_m2m_buf_done(src_buf, run->status ? VB2_BUF_STATE_ERROR :
run               104 drivers/staging/media/imx/imx-media-csc-scaler.c 	v4l2_m2m_buf_done(dst_buf, run->status ? VB2_BUF_STATE_ERROR :
run               108 drivers/staging/media/imx/imx-media-csc-scaler.c 	kfree(run);
run               116 drivers/staging/media/imx/imx-media-csc-scaler.c 	struct ipu_image_convert_run *run;
run               122 drivers/staging/media/imx/imx-media-csc-scaler.c 	run = kzalloc(sizeof(*run), GFP_KERNEL);
run               123 drivers/staging/media/imx/imx-media-csc-scaler.c 	if (!run)
run               126 drivers/staging/media/imx/imx-media-csc-scaler.c 	run->ctx = ctx->icc;
run               127 drivers/staging/media/imx/imx-media-csc-scaler.c 	run->in_phys = vb2_dma_contig_plane_dma_addr(&src_buf->vb2_buf, 0);
run               128 drivers/staging/media/imx/imx-media-csc-scaler.c 	run->out_phys = vb2_dma_contig_plane_dma_addr(&dst_buf->vb2_buf, 0);
run               130 drivers/staging/media/imx/imx-media-csc-scaler.c 	ret = ipu_image_convert_queue(run);
run               120 drivers/staging/media/sunxi/cedrus/cedrus.h 	void (*setup)(struct cedrus_ctx *ctx, struct cedrus_run *run);
run                29 drivers/staging/media/sunxi/cedrus/cedrus_dec.c 	struct cedrus_run run = {};
run                32 drivers/staging/media/sunxi/cedrus/cedrus_dec.c 	run.src = v4l2_m2m_next_src_buf(ctx->fh.m2m_ctx);
run                33 drivers/staging/media/sunxi/cedrus/cedrus_dec.c 	run.dst = v4l2_m2m_next_dst_buf(ctx->fh.m2m_ctx);
run                36 drivers/staging/media/sunxi/cedrus/cedrus_dec.c 	src_req = run.src->vb2_buf.req_obj.req;
run                43 drivers/staging/media/sunxi/cedrus/cedrus_dec.c 		run.mpeg2.slice_params = cedrus_find_control_data(ctx,
run                45 drivers/staging/media/sunxi/cedrus/cedrus_dec.c 		run.mpeg2.quantization = cedrus_find_control_data(ctx,
run                50 drivers/staging/media/sunxi/cedrus/cedrus_dec.c 		run.h264.decode_params = cedrus_find_control_data(ctx,
run                52 drivers/staging/media/sunxi/cedrus/cedrus_dec.c 		run.h264.pps = cedrus_find_control_data(ctx,
run                54 drivers/staging/media/sunxi/cedrus/cedrus_dec.c 		run.h264.scaling_matrix = cedrus_find_control_data(ctx,
run                56 drivers/staging/media/sunxi/cedrus/cedrus_dec.c 		run.h264.slice_params = cedrus_find_control_data(ctx,
run                58 drivers/staging/media/sunxi/cedrus/cedrus_dec.c 		run.h264.sps = cedrus_find_control_data(ctx,
run                66 drivers/staging/media/sunxi/cedrus/cedrus_dec.c 	v4l2_m2m_buf_copy_metadata(run.src, run.dst, true);
run                68 drivers/staging/media/sunxi/cedrus/cedrus_dec.c 	dev->dec_ops[ctx->current_codec]->setup(ctx, &run);
run                93 drivers/staging/media/sunxi/cedrus/cedrus_h264.c 				    struct cedrus_run *run)
run                96 drivers/staging/media/sunxi/cedrus/cedrus_h264.c 	const struct v4l2_ctrl_h264_decode_params *decode = run->h264.decode_params;
run                97 drivers/staging/media/sunxi/cedrus/cedrus_h264.c 	const struct v4l2_ctrl_h264_slice_params *slice = run->h264.slice_params;
run                98 drivers/staging/media/sunxi/cedrus/cedrus_h264.c 	const struct v4l2_ctrl_h264_sps *sps = run->h264.sps;
run               143 drivers/staging/media/sunxi/cedrus/cedrus_h264.c 	output_buf = vb2_to_cedrus_buffer(&run->dst->vb2_buf);
run               167 drivers/staging/media/sunxi/cedrus/cedrus_h264.c 				   struct cedrus_run *run,
run               171 drivers/staging/media/sunxi/cedrus/cedrus_h264.c 	const struct v4l2_ctrl_h264_decode_params *decode = run->h264.decode_params;
run               214 drivers/staging/media/sunxi/cedrus/cedrus_h264.c 				   struct cedrus_run *run)
run               216 drivers/staging/media/sunxi/cedrus/cedrus_h264.c 	const struct v4l2_ctrl_h264_slice_params *slice = run->h264.slice_params;
run               218 drivers/staging/media/sunxi/cedrus/cedrus_h264.c 	_cedrus_write_ref_list(ctx, run,
run               225 drivers/staging/media/sunxi/cedrus/cedrus_h264.c 				   struct cedrus_run *run)
run               227 drivers/staging/media/sunxi/cedrus/cedrus_h264.c 	const struct v4l2_ctrl_h264_slice_params *slice = run->h264.slice_params;
run               229 drivers/staging/media/sunxi/cedrus/cedrus_h264.c 	_cedrus_write_ref_list(ctx, run,
run               236 drivers/staging/media/sunxi/cedrus/cedrus_h264.c 				       struct cedrus_run *run)
run               239 drivers/staging/media/sunxi/cedrus/cedrus_h264.c 		run->h264.scaling_matrix;
run               256 drivers/staging/media/sunxi/cedrus/cedrus_h264.c 					   struct cedrus_run *run)
run               259 drivers/staging/media/sunxi/cedrus/cedrus_h264.c 		run->h264.slice_params;
run               297 drivers/staging/media/sunxi/cedrus/cedrus_h264.c 			      struct cedrus_run *run)
run               299 drivers/staging/media/sunxi/cedrus/cedrus_h264.c 	const struct v4l2_ctrl_h264_decode_params *decode = run->h264.decode_params;
run               300 drivers/staging/media/sunxi/cedrus/cedrus_h264.c 	const struct v4l2_ctrl_h264_slice_params *slice = run->h264.slice_params;
run               301 drivers/staging/media/sunxi/cedrus/cedrus_h264.c 	const struct v4l2_ctrl_h264_pps *pps = run->h264.pps;
run               302 drivers/staging/media/sunxi/cedrus/cedrus_h264.c 	const struct v4l2_ctrl_h264_sps *sps = run->h264.sps;
run               303 drivers/staging/media/sunxi/cedrus/cedrus_h264.c 	struct vb2_buffer *src_buf = &run->src->vb2_buf;
run               335 drivers/staging/media/sunxi/cedrus/cedrus_h264.c 		cedrus_write_pred_weight_table(ctx, run);
run               340 drivers/staging/media/sunxi/cedrus/cedrus_h264.c 		cedrus_write_ref_list0(ctx, run);
run               343 drivers/staging/media/sunxi/cedrus/cedrus_h264.c 		cedrus_write_ref_list1(ctx, run);
run               450 drivers/staging/media/sunxi/cedrus/cedrus_h264.c 			      struct cedrus_run *run)
run               462 drivers/staging/media/sunxi/cedrus/cedrus_h264.c 	cedrus_write_scaling_lists(ctx, run);
run               463 drivers/staging/media/sunxi/cedrus/cedrus_h264.c 	cedrus_write_frame_list(ctx, run);
run               465 drivers/staging/media/sunxi/cedrus/cedrus_h264.c 	cedrus_set_params(ctx, run);
run                75 drivers/staging/media/sunxi/cedrus/cedrus_mpeg2.c static void cedrus_mpeg2_setup(struct cedrus_ctx *ctx, struct cedrus_run *run)
run                92 drivers/staging/media/sunxi/cedrus/cedrus_mpeg2.c 	slice_params = run->mpeg2.slice_params;
run                96 drivers/staging/media/sunxi/cedrus/cedrus_mpeg2.c 	quantization = run->mpeg2.quantization;
run               181 drivers/staging/media/sunxi/cedrus/cedrus_mpeg2.c 	dst_luma_addr = cedrus_dst_buf_addr(ctx, run->dst->vb2_buf.index, 0);
run               182 drivers/staging/media/sunxi/cedrus/cedrus_mpeg2.c 	dst_chroma_addr = cedrus_dst_buf_addr(ctx, run->dst->vb2_buf.index, 1);
run               197 drivers/staging/media/sunxi/cedrus/cedrus_mpeg2.c 	src_buf_addr = vb2_dma_contig_plane_dma_addr(&run->src->vb2_buf, 0);
run               234 drivers/tty/tty_audit.c 		size_t run;
run               236 drivers/tty/tty_audit.c 		run = N_TTY_BUF_SIZE - buf->valid;
run               237 drivers/tty/tty_audit.c 		if (run > size)
run               238 drivers/tty/tty_audit.c 			run = size;
run               239 drivers/tty/tty_audit.c 		memcpy(buf->data + buf->valid, data, run);
run               240 drivers/tty/tty_audit.c 		buf->valid += run;
run               241 drivers/tty/tty_audit.c 		data += run;
run               242 drivers/tty/tty_audit.c 		size -= run;
run               409 drivers/usb/phy/phy-mv-usb.c run:
run               475 drivers/usb/phy/phy-mv-usb.c 		goto run;
run                26 fs/befs/datastream.c 				 befs_blocknr_t blockno, befs_block_run *run);
run                31 fs/befs/datastream.c 				   befs_block_run *run);
run                36 fs/befs/datastream.c 				      befs_block_run *run);
run                53 fs/befs/datastream.c 	befs_block_run run;
run                61 fs/befs/datastream.c 	if (befs_fblock2brun(sb, ds, block, &run) != BEFS_OK) {
run                67 fs/befs/datastream.c 	bh = befs_bread_iaddr(sb, run);
run                96 fs/befs/datastream.c 		 befs_blocknr_t fblock, befs_block_run *run)
run               102 fs/befs/datastream.c 		err = befs_find_brun_direct(sb, data, fblock, run);
run               105 fs/befs/datastream.c 		err = befs_find_brun_indirect(sb, data, fblock, run);
run               108 fs/befs/datastream.c 		err = befs_find_brun_dblindirect(sb, data, fblock, run);
run               252 fs/befs/datastream.c 		      befs_blocknr_t blockno, befs_block_run *run)
run               265 fs/befs/datastream.c 			run->allocation_group = array[i].allocation_group;
run               266 fs/befs/datastream.c 			run->start = array[i].start + offset;
run               267 fs/befs/datastream.c 			run->len = array[i].len - offset;
run               308 fs/befs/datastream.c 			befs_block_run *run)
run               344 fs/befs/datastream.c 				run->allocation_group =
run               346 fs/befs/datastream.c 				run->start =
run               348 fs/befs/datastream.c 				run->len =
run               417 fs/befs/datastream.c 			   befs_block_run *run)
run               509 fs/befs/datastream.c 	*run = fsrun_to_cpu(sb, iaddr_array[block_indx]);
run               517 fs/befs/datastream.c 	run->start += offset;
run               518 fs/befs/datastream.c 	run->len -= offset;
run                12 fs/befs/datastream.h 		     befs_blocknr_t fblock, befs_block_run *run);
run               213 fs/befs/debug.c befs_dump_run(const struct super_block *sb, befs_disk_block_run run)
run               217 fs/befs/debug.c 	befs_block_run n = fsrun_to_cpu(sb, run);
run                74 fs/befs/endian.h 	befs_block_run run;
run                77 fs/befs/endian.h 		run.allocation_group = le32_to_cpu((__force __le32)n.allocation_group);
run                78 fs/befs/endian.h 		run.start = le16_to_cpu((__force __le16)n.start);
run                79 fs/befs/endian.h 		run.len = le16_to_cpu((__force __le16)n.len);
run                81 fs/befs/endian.h 		run.allocation_group = be32_to_cpu((__force __be32)n.allocation_group);
run                82 fs/befs/endian.h 		run.start = be16_to_cpu((__force __be16)n.start);
run                83 fs/befs/endian.h 		run.len = be16_to_cpu((__force __be16)n.len);
run                85 fs/befs/endian.h 	return run;
run                91 fs/befs/endian.h 	befs_disk_block_run run;
run                94 fs/befs/endian.h 		run.allocation_group = cpu_to_le32(n.allocation_group);
run                95 fs/befs/endian.h 		run.start = cpu_to_le16(n.start);
run                96 fs/befs/endian.h 		run.len = cpu_to_le16(n.len);
run                98 fs/befs/endian.h 		run.allocation_group = cpu_to_be32(n.allocation_group);
run                99 fs/befs/endian.h 		run.start = cpu_to_be16(n.start);
run               100 fs/befs/endian.h 		run.len = cpu_to_be16(n.len);
run               102 fs/befs/endian.h 	return run;
run               137 fs/befs/linuxvfs.c 	befs_block_run run = BAD_IADDR;
run               150 fs/befs/linuxvfs.c 	res = befs_fblock2brun(sb, ds, block, &run);
run               159 fs/befs/linuxvfs.c 	disk_off = (ulong) iaddr2blockno(sb, &run);
run              1275 fs/gfs2/dir.c  	int run = 0;
run              1297 fs/gfs2/dir.c  				if (*copied && !run)
run              1299 fs/gfs2/dir.c  				run = 1;
run              1301 fs/gfs2/dir.c  				run = 0;
run               427 fs/jbd2/commit.c 	stats.run.rs_wait = commit_transaction->t_max_wait;
run               428 fs/jbd2/commit.c 	stats.run.rs_request_delay = 0;
run               429 fs/jbd2/commit.c 	stats.run.rs_locked = jiffies;
run               431 fs/jbd2/commit.c 		stats.run.rs_request_delay =
run               433 fs/jbd2/commit.c 				       stats.run.rs_locked);
run               434 fs/jbd2/commit.c 	stats.run.rs_running = jbd2_time_diff(commit_transaction->t_start,
run               435 fs/jbd2/commit.c 					      stats.run.rs_locked);
run               523 fs/jbd2/commit.c 	stats.run.rs_flushing = jiffies;
run               524 fs/jbd2/commit.c 	stats.run.rs_locked = jbd2_time_diff(stats.run.rs_locked,
run               525 fs/jbd2/commit.c 					     stats.run.rs_flushing);
run               560 fs/jbd2/commit.c 	stats.run.rs_logging = jiffies;
run               561 fs/jbd2/commit.c 	stats.run.rs_flushing = jbd2_time_diff(stats.run.rs_flushing,
run               562 fs/jbd2/commit.c 					       stats.run.rs_logging);
run               563 fs/jbd2/commit.c 	stats.run.rs_blocks =
run               565 fs/jbd2/commit.c 	stats.run.rs_blocks_logged = 0;
run               816 fs/jbd2/commit.c 		stats.run.rs_blocks_logged++;
run               862 fs/jbd2/commit.c 		stats.run.rs_blocks_logged++;
run               884 fs/jbd2/commit.c 	stats.run.rs_blocks_logged++;
run              1085 fs/jbd2/commit.c 	stats.run.rs_logging = jbd2_time_diff(stats.run.rs_logging,
run              1092 fs/jbd2/commit.c 	stats.run.rs_handle_count =
run              1095 fs/jbd2/commit.c 			     commit_transaction->t_tid, &stats.run);
run              1142 fs/jbd2/commit.c 	journal->j_stats.run.rs_wait += stats.run.rs_wait;
run              1143 fs/jbd2/commit.c 	journal->j_stats.run.rs_request_delay += stats.run.rs_request_delay;
run              1144 fs/jbd2/commit.c 	journal->j_stats.run.rs_running += stats.run.rs_running;
run              1145 fs/jbd2/commit.c 	journal->j_stats.run.rs_locked += stats.run.rs_locked;
run              1146 fs/jbd2/commit.c 	journal->j_stats.run.rs_flushing += stats.run.rs_flushing;
run              1147 fs/jbd2/commit.c 	journal->j_stats.run.rs_logging += stats.run.rs_logging;
run              1148 fs/jbd2/commit.c 	journal->j_stats.run.rs_handle_count += stats.run.rs_handle_count;
run              1149 fs/jbd2/commit.c 	journal->j_stats.run.rs_blocks += stats.run.rs_blocks;
run              1150 fs/jbd2/commit.c 	journal->j_stats.run.rs_blocks_logged += stats.run.rs_blocks_logged;
run              1001 fs/jbd2/journal.c 	    jiffies_to_msecs(s->stats->run.rs_wait / s->stats->ts_tid));
run              1004 fs/jbd2/journal.c 	    jiffies_to_msecs(s->stats->run.rs_request_delay /
run              1007 fs/jbd2/journal.c 	    jiffies_to_msecs(s->stats->run.rs_running / s->stats->ts_tid));
run              1009 fs/jbd2/journal.c 	    jiffies_to_msecs(s->stats->run.rs_locked / s->stats->ts_tid));
run              1011 fs/jbd2/journal.c 	    jiffies_to_msecs(s->stats->run.rs_flushing / s->stats->ts_tid));
run              1013 fs/jbd2/journal.c 	    jiffies_to_msecs(s->stats->run.rs_logging / s->stats->ts_tid));
run              1017 fs/jbd2/journal.c 	    s->stats->run.rs_handle_count / s->stats->ts_tid);
run              1019 fs/jbd2/journal.c 	    s->stats->run.rs_blocks / s->stats->ts_tid);
run              1021 fs/jbd2/journal.c 	    s->stats->run.rs_blocks_logged / s->stats->ts_tid);
run               139 fs/omfs/bitmap.c 	int i, run, bit;
run               151 fs/omfs/bitmap.c 			run = count_run(&sbi->s_imap[i], bits_per_entry,
run               154 fs/omfs/bitmap.c 			if (run >= min_request)
run               156 fs/omfs/bitmap.c 			bit += run;
run               164 fs/omfs/bitmap.c 	*return_size = run;
run               165 fs/omfs/bitmap.c 	ret = set_run(sb, i, bits_per_entry, bit, run, 1);
run                52 include/linux/blk-mq.h 	unsigned long		run;
run               743 include/linux/jbd2.h 	struct transaction_run_stats_s run;
run               279 include/linux/kvm_host.h 	struct kvm_run *run;
run               651 include/linux/phy.h 	int (*run)(struct phy_device *phydev);
run              1172 include/linux/phy.h 		       int (*run)(struct phy_device *));
run              1174 include/linux/phy.h 			      int (*run)(struct phy_device *));
run              1176 include/linux/phy.h 			       int (*run)(struct phy_device *));
run                40 include/video/imx-ipu-image-convert.h typedef void (*ipu_image_convert_cb_t)(struct ipu_image_convert_run *run,
run               139 include/video/imx-ipu-image-convert.h int ipu_image_convert_queue(struct ipu_image_convert_run *run);
run               489 kernel/stop_machine.c 	int run;
run               492 kernel/stop_machine.c 	run = !list_empty(&stopper->works);
run               494 kernel/stop_machine.c 	return run;
run              2946 mm/ksm.c       KSM_ATTR(run);
run               900 sound/soc/qcom/qdsp6/q6asm.c 	struct asm_session_cmd_run_v2 *run;
run               905 sound/soc/qcom/qdsp6/q6asm.c 	pkt_size = APR_HDR_SIZE + sizeof(*run);
run               911 sound/soc/qcom/qdsp6/q6asm.c 	run = p + APR_HDR_SIZE;
run               916 sound/soc/qcom/qdsp6/q6asm.c 	run->flags = flags;
run               917 sound/soc/qcom/qdsp6/q6asm.c 	run->time_lsw = lsw_ts;
run               918 sound/soc/qcom/qdsp6/q6asm.c 	run->time_msw = msw_ts;
run               491 sound/soc/sof/core.c 	if (!sof_ops(sdev) || !sof_ops(sdev)->probe || !sof_ops(sdev)->run ||
run               370 sound/soc/sof/imx/imx8.c 	.run		= imx8_run,
run                74 sound/soc/sof/intel/apl.c 	.run = hda_dsp_cl_boot_firmware,
run               535 sound/soc/sof/intel/bdw.c 	.run            = bdw_run,
run               471 sound/soc/sof/intel/byt.c 	.run		= byt_run,
run               632 sound/soc/sof/intel/byt.c 	.run		= byt_run,
run               691 sound/soc/sof/intel/byt.c 	.run		= byt_run,
run               240 sound/soc/sof/intel/cnl.c 	.run = hda_dsp_cl_boot_firmware,
run               253 sound/soc/sof/intel/hda-stream.c 	u32 run;
run               273 sound/soc/sof/intel/hda-stream.c 					sd_offset, run,
run               274 sound/soc/sof/intel/hda-stream.c 					((run &	dma_start) == dma_start),
run               292 sound/soc/sof/intel/hda-stream.c 						sd_offset, run,
run               293 sound/soc/sof/intel/hda-stream.c 						!(run &	dma_start),
run               331 sound/soc/sof/intel/hda-stream.c 	u32 run;
run               354 sound/soc/sof/intel/hda-stream.c 					    sd_offset, run,
run               355 sound/soc/sof/intel/hda-stream.c 					    !(run & dma_start),
run               416 sound/soc/sof/intel/hda-stream.c 					    sd_offset, run,
run               417 sound/soc/sof/intel/hda-stream.c 					    !(run & dma_start),
run                48 sound/soc/sof/ops.h 	return sof_ops(sdev)->run(sdev);
run                85 sound/soc/sof/sof-priv.h 	int (*run)(struct snd_sof_dev *sof_dev); /* mandatory */
run                36 tools/firewire/nosy-dump.c static int run = 1;
run               127 tools/firewire/nosy-dump.c 	if (run == 1) {
run               128 tools/firewire/nosy-dump.c 		run = 0;
run               973 tools/firewire/nosy-dump.c 	while (run) {
run              1946 tools/perf/builtin-script.c 				counts->run,
run               257 tools/perf/builtin-stat.c 		count->ena = count->run = val;
run               298 tools/perf/builtin-stat.c 				perf_counts(counter->counts, cpu, thread)->run = 0;
run               316 tools/perf/builtin-stat.c 						count->val, count->ena, count->run);
run               296 tools/perf/lib/include/perf/event.h 			__u64	 run;
run                18 tools/perf/lib/include/perf/evsel.h 			uint64_t run;
run               202 tools/perf/tests/event-times.c 	err = !(count.ena == count.run);
run               206 tools/perf/tests/event-times.c 		 count.ena, count.run);
run                76 tools/perf/tests/stat.c 	TEST_ASSERT_VAL("wrong ena",    st->run    == 300);
run                86 tools/perf/tests/stat.c 	count.run = 300;
run               161 tools/perf/util/c++/clang.cpp 	PM.run(*Module);
run              1284 tools/perf/util/evsel.c 	count->run = count->run - tmp.run;
run              1293 tools/perf/util/evsel.c 		if (count->run == 0) {
run              1296 tools/perf/util/evsel.c 		} else if (count->run < count->ena) {
run              1298 tools/perf/util/evsel.c 			count->val = (u64)((double) count->val * count->ena / count->run);
run              1316 tools/perf/util/evsel.c 		      u64 val, u64 ena, u64 run)
run              1324 tools/perf/util/evsel.c 	count->run    = run;
run              1335 tools/perf/util/evsel.c 	u64 nr, ena = 0, run = 0, i;
run              1346 tools/perf/util/evsel.c 		run = *data++;
run              1351 tools/perf/util/evsel.c 			      v[0].value, ena, run);
run              1361 tools/perf/util/evsel.c 				      v[i].value, ena, run);
run              1382 tools/perf/util/scripting-engines/trace-event-python.c 	tuple_set_u64(t, n++, count->run);
run               899 tools/perf/util/session.c 	event->stat.run    = bswap_64(event->stat.run);
run                24 tools/perf/util/stat-display.c 			  u64 run, u64 ena)
run                29 tools/perf/util/stat-display.c 					run,
run                31 tools/perf/util/stat-display.c 					ena ? 100.0 * run / ena : 100.0);
run                32 tools/perf/util/stat-display.c 	} else if (run != ena) {
run                33 tools/perf/util/stat-display.c 		fprintf(config->output, "  (%.2f%%)", 100.0 * run / ena);
run               393 tools/perf/util/stat-display.c 		     char *prefix, u64 run, u64 ena, double noise,
run               433 tools/perf/util/stat-display.c 	if (run == 0 || ena == 0 || counter->counts->scaled == -1) {
run               464 tools/perf/util/stat-display.c 		print_running(config, run, ena);
run               480 tools/perf/util/stat-display.c 		print_running(config, run, ena);
run               488 tools/perf/util/stat-display.c 		print_running(config, run, ena);
run               582 tools/perf/util/stat-display.c 	u64 ena, run, val;
run               607 tools/perf/util/stat-display.c 		if (counts->ena == 0 || counts->run == 0 ||
run               610 tools/perf/util/stat-display.c 			ad->run = 0;
run               615 tools/perf/util/stat-display.c 		ad->run += counts->run;
run               626 tools/perf/util/stat-display.c 	u64 ena, run, val;
run               631 tools/perf/util/stat-display.c 	ad.val = ad.ena = ad.run = 0;
run               638 tools/perf/util/stat-display.c 	run = ad.run;
run               649 tools/perf/util/stat-display.c 		 run, ena, 1.0, &rt_stat);
run               709 tools/perf/util/stat-display.c 		u64 ena = 0, run = 0, val = 0;
run               714 tools/perf/util/stat-display.c 			run += perf_counts(counter->counts, cpu, thread)->run;
run               730 tools/perf/util/stat-display.c 		buf[i].run = run;
run               766 tools/perf/util/stat-display.c 				 prefix, buf[thread].run, buf[thread].ena, 1.0,
run               770 tools/perf/util/stat-display.c 				 prefix, buf[thread].run, buf[thread].ena, 1.0,
run               827 tools/perf/util/stat-display.c 	ad->run += perf_counts(counter->counts, ad->cpu, 0)->run;
run               838 tools/perf/util/stat-display.c 	u64 ena, run, val;
run               849 tools/perf/util/stat-display.c 		run = ad.run;
run               855 tools/perf/util/stat-display.c 		printout(config, cpu, 0, counter, uval, prefix, run, ena, 1.0,
run               869 tools/perf/util/stat-display.c 	u64 ena, run, val;
run               885 tools/perf/util/stat-display.c 			run = perf_counts(counter->counts, cpu, 0)->run;
run               888 tools/perf/util/stat-display.c 			printout(config, cpu, 0, counter, uval, prefix, run, ena, 1.0,
run              1062 tools/perf/util/stat-display.c 		double run = (double) config->walltime_run[idx] / NSEC_PER_SEC;
run              1063 tools/perf/util/stat-display.c 		int h, n = 1 + abs((int) (100.0 * (run - avg)/run) / 5);
run              1066 tools/perf/util/stat-display.c 			precision, run, precision, run - avg);
run               170 tools/perf/util/stat.c 		evsel->prev_raw_counts->aggr.run = 0;
run               269 tools/perf/util/stat.c 	if (!(vals->run && vals->ena))
run               323 tools/perf/util/stat.c 		aggr->run += count->run;
run               361 tools/perf/util/stat.c 	aggr->val = aggr->ena = aggr->run = 0;
run               412 tools/perf/util/stat.c 	count.run = st->run;
run               433 tools/perf/util/stat.c 		       st->val, st->ena, st->run);
run               153 tools/perf/util/stat.h 	u64 run;
run              1105 tools/perf/util/synthetic-events.c 	event.run       = count->run;
run               799 tools/testing/selftests/bpf/test_sockmap.c 		goto run;
run               827 tools/testing/selftests/bpf/test_sockmap.c run:
run               154 tools/testing/selftests/kvm/dirty_log_test.c 	struct kvm_run *run;
run               156 tools/testing/selftests/kvm/dirty_log_test.c 	run = vcpu_state(vm, VCPU_ID);
run               172 tools/testing/selftests/kvm/dirty_log_test.c 				    exit_reason_str(run->exit_reason));
run                94 tools/testing/selftests/kvm/lib/aarch64/ucall.c 	struct kvm_run *run = vcpu_state(vm, vcpu_id);
run                97 tools/testing/selftests/kvm/lib/aarch64/ucall.c 	if (run->exit_reason == KVM_EXIT_MMIO &&
run                98 tools/testing/selftests/kvm/lib/aarch64/ucall.c 	    run->mmio.phys_addr == (uint64_t)ucall_exit_mmio_addr) {
run               101 tools/testing/selftests/kvm/lib/aarch64/ucall.c 		TEST_ASSERT(run->mmio.is_write && run->mmio.len == 8,
run               103 tools/testing/selftests/kvm/lib/aarch64/ucall.c 		memcpy(&gva, run->mmio.data, sizeof(gva));
run               247 tools/testing/selftests/kvm/lib/s390x/processor.c 	struct kvm_run *run;
run               267 tools/testing/selftests/kvm/lib/s390x/processor.c 	run = vcpu_state(vm, vcpuid);
run               268 tools/testing/selftests/kvm/lib/s390x/processor.c 	run->psw_mask = 0x0400000180000000ULL;  /* DAT enabled + 64 bit mode */
run               269 tools/testing/selftests/kvm/lib/s390x/processor.c 	run->psw_addr = (uintptr_t)guest_code;
run                38 tools/testing/selftests/kvm/lib/s390x/ucall.c 	struct kvm_run *run = vcpu_state(vm, vcpu_id);
run                41 tools/testing/selftests/kvm/lib/s390x/ucall.c 	if (run->exit_reason == KVM_EXIT_S390_SIEIC &&
run                42 tools/testing/selftests/kvm/lib/s390x/ucall.c 	    run->s390_sieic.icptcode == 4 &&
run                43 tools/testing/selftests/kvm/lib/s390x/ucall.c 	    (run->s390_sieic.ipa >> 8) == 0x83 &&    /* 0x83 means DIAGNOSE */
run                44 tools/testing/selftests/kvm/lib/s390x/ucall.c 	    (run->s390_sieic.ipb >> 16) == 0x501) {
run                45 tools/testing/selftests/kvm/lib/s390x/ucall.c 		int reg = run->s390_sieic.ipa & 0xf;
run                47 tools/testing/selftests/kvm/lib/s390x/ucall.c 		memcpy(&ucall, addr_gva2hva(vm, run->s.regs.gprs[reg]),
run                40 tools/testing/selftests/kvm/lib/x86_64/ucall.c 	struct kvm_run *run = vcpu_state(vm, vcpu_id);
run                43 tools/testing/selftests/kvm/lib/x86_64/ucall.c 	if (run->exit_reason == KVM_EXIT_IO && run->io.port == UCALL_PIO_PORT) {
run                35 tools/testing/selftests/kvm/s390x/memop.c 	struct kvm_run *run;
run                51 tools/testing/selftests/kvm/s390x/memop.c 	run = vcpu_state(vm, VCPU_ID);
run                67 tools/testing/selftests/kvm/s390x/memop.c 	TEST_ASSERT(run->exit_reason == KVM_EXIT_S390_SIEIC,
run                69 tools/testing/selftests/kvm/s390x/memop.c 		    run->exit_reason,
run                70 tools/testing/selftests/kvm/s390x/memop.c 		    exit_reason_str(run->exit_reason));
run               149 tools/testing/selftests/kvm/s390x/memop.c 	run->psw_mask &= ~(3UL << (63 - 17));
run               150 tools/testing/selftests/kvm/s390x/memop.c 	run->psw_mask |= 1UL << (63 - 17);  /* Enable AR mode */
run               160 tools/testing/selftests/kvm/s390x/memop.c 	run->psw_mask &= ~(3UL << (63 - 17));   /* Disable AR mode */
run                69 tools/testing/selftests/kvm/s390x/sync_regs_test.c 	struct kvm_run *run;
run                86 tools/testing/selftests/kvm/s390x/sync_regs_test.c 	run = vcpu_state(vm, VCPU_ID);
run                89 tools/testing/selftests/kvm/s390x/sync_regs_test.c 	run->kvm_valid_regs = INVALID_SYNC_FIELD;
run                96 tools/testing/selftests/kvm/s390x/sync_regs_test.c 	run->kvm_valid_regs = INVALID_SYNC_FIELD | TEST_SYNC_FIELDS;
run               104 tools/testing/selftests/kvm/s390x/sync_regs_test.c 	run->kvm_dirty_regs = INVALID_SYNC_FIELD;
run               111 tools/testing/selftests/kvm/s390x/sync_regs_test.c 	run->kvm_dirty_regs = INVALID_SYNC_FIELD | TEST_SYNC_FIELDS;
run               119 tools/testing/selftests/kvm/s390x/sync_regs_test.c 	run->kvm_valid_regs = TEST_SYNC_FIELDS;
run               122 tools/testing/selftests/kvm/s390x/sync_regs_test.c 	TEST_ASSERT(run->exit_reason == KVM_EXIT_S390_SIEIC,
run               124 tools/testing/selftests/kvm/s390x/sync_regs_test.c 		    run->exit_reason,
run               125 tools/testing/selftests/kvm/s390x/sync_regs_test.c 		    exit_reason_str(run->exit_reason));
run               126 tools/testing/selftests/kvm/s390x/sync_regs_test.c 	TEST_ASSERT(run->s390_sieic.icptcode == 4 &&
run               127 tools/testing/selftests/kvm/s390x/sync_regs_test.c 		    (run->s390_sieic.ipa >> 8) == 0x83 &&
run               128 tools/testing/selftests/kvm/s390x/sync_regs_test.c 		    (run->s390_sieic.ipb >> 16) == 0x501,
run               130 tools/testing/selftests/kvm/s390x/sync_regs_test.c 		    run->s390_sieic.icptcode, run->s390_sieic.ipa,
run               131 tools/testing/selftests/kvm/s390x/sync_regs_test.c 		    run->s390_sieic.ipb);
run               134 tools/testing/selftests/kvm/s390x/sync_regs_test.c 	compare_regs(&regs, &run->s.regs);
run               137 tools/testing/selftests/kvm/s390x/sync_regs_test.c 	compare_sregs(&sregs, &run->s.regs);
run               140 tools/testing/selftests/kvm/s390x/sync_regs_test.c 	run->s.regs.gprs[11] = 0xBAD1DEA;
run               141 tools/testing/selftests/kvm/s390x/sync_regs_test.c 	run->s.regs.acrs[0] = 1 << 11;
run               143 tools/testing/selftests/kvm/s390x/sync_regs_test.c 	run->kvm_valid_regs = TEST_SYNC_FIELDS;
run               144 tools/testing/selftests/kvm/s390x/sync_regs_test.c 	run->kvm_dirty_regs = KVM_SYNC_GPRS | KVM_SYNC_ACRS;
run               147 tools/testing/selftests/kvm/s390x/sync_regs_test.c 	TEST_ASSERT(run->exit_reason == KVM_EXIT_S390_SIEIC,
run               149 tools/testing/selftests/kvm/s390x/sync_regs_test.c 		    run->exit_reason,
run               150 tools/testing/selftests/kvm/s390x/sync_regs_test.c 		    exit_reason_str(run->exit_reason));
run               151 tools/testing/selftests/kvm/s390x/sync_regs_test.c 	TEST_ASSERT(run->s.regs.gprs[11] == 0xBAD1DEA + 1,
run               153 tools/testing/selftests/kvm/s390x/sync_regs_test.c 		    run->s.regs.gprs[11]);
run               154 tools/testing/selftests/kvm/s390x/sync_regs_test.c 	TEST_ASSERT(run->s.regs.acrs[0]  == 1 << 11,
run               156 tools/testing/selftests/kvm/s390x/sync_regs_test.c 		    run->s.regs.acrs[0]);
run               159 tools/testing/selftests/kvm/s390x/sync_regs_test.c 	compare_regs(&regs, &run->s.regs);
run               162 tools/testing/selftests/kvm/s390x/sync_regs_test.c 	compare_sregs(&sregs, &run->s.regs);
run               167 tools/testing/selftests/kvm/s390x/sync_regs_test.c 	run->kvm_valid_regs = TEST_SYNC_FIELDS;
run               168 tools/testing/selftests/kvm/s390x/sync_regs_test.c 	run->kvm_dirty_regs = 0;
run               169 tools/testing/selftests/kvm/s390x/sync_regs_test.c 	run->s.regs.gprs[11] = 0xDEADBEEF;
run               172 tools/testing/selftests/kvm/s390x/sync_regs_test.c 	TEST_ASSERT(run->exit_reason == KVM_EXIT_S390_SIEIC,
run               174 tools/testing/selftests/kvm/s390x/sync_regs_test.c 		    run->exit_reason,
run               175 tools/testing/selftests/kvm/s390x/sync_regs_test.c 		    exit_reason_str(run->exit_reason));
run               176 tools/testing/selftests/kvm/s390x/sync_regs_test.c 	TEST_ASSERT(run->s.regs.gprs[11] != 0xDEADBEEF,
run               178 tools/testing/selftests/kvm/s390x/sync_regs_test.c 		    run->s.regs.gprs[11]);
run                66 tools/testing/selftests/kvm/x86_64/cr4_cpuid_sync_test.c 	struct kvm_run *run;
run                85 tools/testing/selftests/kvm/x86_64/cr4_cpuid_sync_test.c 	run = vcpu_state(vm, VCPU_ID);
run                91 tools/testing/selftests/kvm/x86_64/cr4_cpuid_sync_test.c 		TEST_ASSERT(run->exit_reason == KVM_EXIT_IO,
run                93 tools/testing/selftests/kvm/x86_64/cr4_cpuid_sync_test.c 			    run->exit_reason,
run                94 tools/testing/selftests/kvm/x86_64/cr4_cpuid_sync_test.c 			    exit_reason_str(run->exit_reason));
run                78 tools/testing/selftests/kvm/x86_64/evmcs_test.c 	struct kvm_run *run;
run                96 tools/testing/selftests/kvm/x86_64/evmcs_test.c 	run = vcpu_state(vm, VCPU_ID);
run               105 tools/testing/selftests/kvm/x86_64/evmcs_test.c 		TEST_ASSERT(run->exit_reason == KVM_EXIT_IO,
run               107 tools/testing/selftests/kvm/x86_64/evmcs_test.c 			    stage, run->exit_reason,
run               108 tools/testing/selftests/kvm/x86_64/evmcs_test.c 			    exit_reason_str(run->exit_reason));
run               140 tools/testing/selftests/kvm/x86_64/evmcs_test.c 		run = vcpu_state(vm, VCPU_ID);
run                36 tools/testing/selftests/kvm/x86_64/mmio_warning_test.c 	struct kvm_run *run;
run                44 tools/testing/selftests/kvm/x86_64/mmio_warning_test.c 	struct kvm_run *run = tc->run;
run                48 tools/testing/selftests/kvm/x86_64/mmio_warning_test.c 		res, run->exit_reason, run->internal.suberror);
run                57 tools/testing/selftests/kvm/x86_64/mmio_warning_test.c 	struct kvm_run *run;
run                66 tools/testing/selftests/kvm/x86_64/mmio_warning_test.c 	run = (struct kvm_run *)mmap(0, 4096, PROT_READ|PROT_WRITE, MAP_SHARED,
run                69 tools/testing/selftests/kvm/x86_64/mmio_warning_test.c 	tc.run = run;
run                50 tools/testing/selftests/kvm/x86_64/platform_info_test.c 	struct kvm_run *run = vcpu_state(vm, VCPU_ID);
run                55 tools/testing/selftests/kvm/x86_64/platform_info_test.c 	TEST_ASSERT(run->exit_reason == KVM_EXIT_IO,
run                57 tools/testing/selftests/kvm/x86_64/platform_info_test.c 			run->exit_reason,
run                58 tools/testing/selftests/kvm/x86_64/platform_info_test.c 			exit_reason_str(run->exit_reason));
run                71 tools/testing/selftests/kvm/x86_64/platform_info_test.c 	struct kvm_run *run = vcpu_state(vm, VCPU_ID);
run                75 tools/testing/selftests/kvm/x86_64/platform_info_test.c 	TEST_ASSERT(run->exit_reason == KVM_EXIT_SHUTDOWN,
run                77 tools/testing/selftests/kvm/x86_64/platform_info_test.c 			run->exit_reason,
run                78 tools/testing/selftests/kvm/x86_64/platform_info_test.c 			exit_reason_str(run->exit_reason));
run                94 tools/testing/selftests/kvm/x86_64/smm_test.c 	struct kvm_run *run;
run               103 tools/testing/selftests/kvm/x86_64/smm_test.c 	run = vcpu_state(vm, VCPU_ID);
run               126 tools/testing/selftests/kvm/x86_64/smm_test.c 		TEST_ASSERT(run->exit_reason == KVM_EXIT_IO,
run               128 tools/testing/selftests/kvm/x86_64/smm_test.c 			    stage, run->exit_reason,
run               129 tools/testing/selftests/kvm/x86_64/smm_test.c 			    exit_reason_str(run->exit_reason));
run               150 tools/testing/selftests/kvm/x86_64/smm_test.c 		run = vcpu_state(vm, VCPU_ID);
run               126 tools/testing/selftests/kvm/x86_64/state_test.c 	struct kvm_run *run;
run               134 tools/testing/selftests/kvm/x86_64/state_test.c 	run = vcpu_state(vm, VCPU_ID);
run               148 tools/testing/selftests/kvm/x86_64/state_test.c 		TEST_ASSERT(run->exit_reason == KVM_EXIT_IO,
run               150 tools/testing/selftests/kvm/x86_64/state_test.c 			    stage, run->exit_reason,
run               151 tools/testing/selftests/kvm/x86_64/state_test.c 			    exit_reason_str(run->exit_reason));
run               182 tools/testing/selftests/kvm/x86_64/state_test.c 		run = vcpu_state(vm, VCPU_ID);
run                83 tools/testing/selftests/kvm/x86_64/sync_regs_test.c 	struct kvm_run *run;
run               105 tools/testing/selftests/kvm/x86_64/sync_regs_test.c 	run = vcpu_state(vm, VCPU_ID);
run               108 tools/testing/selftests/kvm/x86_64/sync_regs_test.c 	run->kvm_valid_regs = INVALID_SYNC_FIELD;
run               115 tools/testing/selftests/kvm/x86_64/sync_regs_test.c 	run->kvm_valid_regs = INVALID_SYNC_FIELD | TEST_SYNC_FIELDS;
run               123 tools/testing/selftests/kvm/x86_64/sync_regs_test.c 	run->kvm_dirty_regs = INVALID_SYNC_FIELD;
run               130 tools/testing/selftests/kvm/x86_64/sync_regs_test.c 	run->kvm_dirty_regs = INVALID_SYNC_FIELD | TEST_SYNC_FIELDS;
run               139 tools/testing/selftests/kvm/x86_64/sync_regs_test.c 	run->kvm_valid_regs = TEST_SYNC_FIELDS;
run               141 tools/testing/selftests/kvm/x86_64/sync_regs_test.c 	TEST_ASSERT(run->exit_reason == KVM_EXIT_IO,
run               143 tools/testing/selftests/kvm/x86_64/sync_regs_test.c 		    run->exit_reason,
run               144 tools/testing/selftests/kvm/x86_64/sync_regs_test.c 		    exit_reason_str(run->exit_reason));
run               147 tools/testing/selftests/kvm/x86_64/sync_regs_test.c 	compare_regs(&regs, &run->s.regs.regs);
run               150 tools/testing/selftests/kvm/x86_64/sync_regs_test.c 	compare_sregs(&sregs, &run->s.regs.sregs);
run               153 tools/testing/selftests/kvm/x86_64/sync_regs_test.c 	compare_vcpu_events(&events, &run->s.regs.events);
run               156 tools/testing/selftests/kvm/x86_64/sync_regs_test.c 	run->s.regs.regs.rbx = 0xBAD1DEA;
run               157 tools/testing/selftests/kvm/x86_64/sync_regs_test.c 	run->s.regs.sregs.apic_base = 1 << 11;
run               160 tools/testing/selftests/kvm/x86_64/sync_regs_test.c 	run->kvm_valid_regs = TEST_SYNC_FIELDS;
run               161 tools/testing/selftests/kvm/x86_64/sync_regs_test.c 	run->kvm_dirty_regs = KVM_SYNC_X86_REGS | KVM_SYNC_X86_SREGS;
run               163 tools/testing/selftests/kvm/x86_64/sync_regs_test.c 	TEST_ASSERT(run->exit_reason == KVM_EXIT_IO,
run               165 tools/testing/selftests/kvm/x86_64/sync_regs_test.c 		    run->exit_reason,
run               166 tools/testing/selftests/kvm/x86_64/sync_regs_test.c 		    exit_reason_str(run->exit_reason));
run               167 tools/testing/selftests/kvm/x86_64/sync_regs_test.c 	TEST_ASSERT(run->s.regs.regs.rbx == 0xBAD1DEA + 1,
run               169 tools/testing/selftests/kvm/x86_64/sync_regs_test.c 		    run->s.regs.regs.rbx);
run               170 tools/testing/selftests/kvm/x86_64/sync_regs_test.c 	TEST_ASSERT(run->s.regs.sregs.apic_base == 1 << 11,
run               172 tools/testing/selftests/kvm/x86_64/sync_regs_test.c 		    run->s.regs.sregs.apic_base);
run               175 tools/testing/selftests/kvm/x86_64/sync_regs_test.c 	compare_regs(&regs, &run->s.regs.regs);
run               178 tools/testing/selftests/kvm/x86_64/sync_regs_test.c 	compare_sregs(&sregs, &run->s.regs.sregs);
run               181 tools/testing/selftests/kvm/x86_64/sync_regs_test.c 	compare_vcpu_events(&events, &run->s.regs.events);
run               186 tools/testing/selftests/kvm/x86_64/sync_regs_test.c 	run->kvm_valid_regs = TEST_SYNC_FIELDS;
run               187 tools/testing/selftests/kvm/x86_64/sync_regs_test.c 	run->kvm_dirty_regs = 0;
run               188 tools/testing/selftests/kvm/x86_64/sync_regs_test.c 	run->s.regs.regs.rbx = 0xDEADBEEF;
run               190 tools/testing/selftests/kvm/x86_64/sync_regs_test.c 	TEST_ASSERT(run->exit_reason == KVM_EXIT_IO,
run               192 tools/testing/selftests/kvm/x86_64/sync_regs_test.c 		    run->exit_reason,
run               193 tools/testing/selftests/kvm/x86_64/sync_regs_test.c 		    exit_reason_str(run->exit_reason));
run               194 tools/testing/selftests/kvm/x86_64/sync_regs_test.c 	TEST_ASSERT(run->s.regs.regs.rbx != 0xDEADBEEF,
run               196 tools/testing/selftests/kvm/x86_64/sync_regs_test.c 		    run->s.regs.regs.rbx);
run               202 tools/testing/selftests/kvm/x86_64/sync_regs_test.c 	run->kvm_valid_regs = 0;
run               203 tools/testing/selftests/kvm/x86_64/sync_regs_test.c 	run->kvm_dirty_regs = 0;
run               204 tools/testing/selftests/kvm/x86_64/sync_regs_test.c 	run->s.regs.regs.rbx = 0xAAAA;
run               208 tools/testing/selftests/kvm/x86_64/sync_regs_test.c 	TEST_ASSERT(run->exit_reason == KVM_EXIT_IO,
run               210 tools/testing/selftests/kvm/x86_64/sync_regs_test.c 		    run->exit_reason,
run               211 tools/testing/selftests/kvm/x86_64/sync_regs_test.c 		    exit_reason_str(run->exit_reason));
run               212 tools/testing/selftests/kvm/x86_64/sync_regs_test.c 	TEST_ASSERT(run->s.regs.regs.rbx == 0xAAAA,
run               214 tools/testing/selftests/kvm/x86_64/sync_regs_test.c 		    run->s.regs.regs.rbx);
run               224 tools/testing/selftests/kvm/x86_64/sync_regs_test.c 	run->kvm_valid_regs = 0;
run               225 tools/testing/selftests/kvm/x86_64/sync_regs_test.c 	run->kvm_dirty_regs = TEST_SYNC_FIELDS;
run               226 tools/testing/selftests/kvm/x86_64/sync_regs_test.c 	run->s.regs.regs.rbx = 0xBBBB;
run               228 tools/testing/selftests/kvm/x86_64/sync_regs_test.c 	TEST_ASSERT(run->exit_reason == KVM_EXIT_IO,
run               230 tools/testing/selftests/kvm/x86_64/sync_regs_test.c 		    run->exit_reason,
run               231 tools/testing/selftests/kvm/x86_64/sync_regs_test.c 		    exit_reason_str(run->exit_reason));
run               232 tools/testing/selftests/kvm/x86_64/sync_regs_test.c 	TEST_ASSERT(run->s.regs.regs.rbx == 0xBBBB,
run               234 tools/testing/selftests/kvm/x86_64/sync_regs_test.c 		    run->s.regs.regs.rbx);
run                67 tools/testing/selftests/kvm/x86_64/vmx_close_while_nested_test.c 		volatile struct kvm_run *run = vcpu_state(vm, VCPU_ID);
run                71 tools/testing/selftests/kvm/x86_64/vmx_close_while_nested_test.c 		TEST_ASSERT(run->exit_reason == KVM_EXIT_IO,
run                73 tools/testing/selftests/kvm/x86_64/vmx_close_while_nested_test.c 			    run->exit_reason,
run                74 tools/testing/selftests/kvm/x86_64/vmx_close_while_nested_test.c 			    exit_reason_str(run->exit_reason));
run                76 tools/testing/selftests/kvm/x86_64/vmx_close_while_nested_test.c 		if (run->io.port == PORT_L0_EXIT)
run                77 tools/testing/selftests/kvm/x86_64/vmx_dirty_log_test.c 	struct kvm_run *run;
run                88 tools/testing/selftests/kvm/x86_64/vmx_dirty_log_test.c 	run = vcpu_state(vm, VCPU_ID);
run               122 tools/testing/selftests/kvm/x86_64/vmx_dirty_log_test.c 		TEST_ASSERT(run->exit_reason == KVM_EXIT_IO,
run               124 tools/testing/selftests/kvm/x86_64/vmx_dirty_log_test.c 			    run->exit_reason,
run               125 tools/testing/selftests/kvm/x86_64/vmx_dirty_log_test.c 			    exit_reason_str(run->exit_reason));
run               142 tools/testing/selftests/kvm/x86_64/vmx_tsc_adjust_test.c 		volatile struct kvm_run *run = vcpu_state(vm, VCPU_ID);
run               146 tools/testing/selftests/kvm/x86_64/vmx_tsc_adjust_test.c 		TEST_ASSERT(run->exit_reason == KVM_EXIT_IO,
run               148 tools/testing/selftests/kvm/x86_64/vmx_tsc_adjust_test.c 			    run->exit_reason,
run               149 tools/testing/selftests/kvm/x86_64/vmx_tsc_adjust_test.c 			    exit_reason_str(run->exit_reason));
run               659 tools/testing/vsock/vsock_diag_test.c 		void (*run)(unsigned int peer_cid);
run               665 tools/testing/vsock/vsock_diag_test.c 			run = test_cases[i].run_client;
run               667 tools/testing/vsock/vsock_diag_test.c 			run = test_cases[i].run_server;
run               669 tools/testing/vsock/vsock_diag_test.c 		if (run)
run               670 tools/testing/vsock/vsock_diag_test.c 			run(peer_cid);
run               283 virt/kvm/arm/arch_timer.c 	struct kvm_sync_regs *regs = &vcpu->run->s.regs;
run               557 virt/kvm/arm/arch_timer.c 	struct kvm_sync_regs *sregs = &vcpu->run->s.regs;
run               682 virt/kvm/arm/arm.c int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu, struct kvm_run *run)
run               693 virt/kvm/arm/arm.c 	if (run->exit_reason == KVM_EXIT_MMIO) {
run               694 virt/kvm/arm/arm.c 		ret = kvm_handle_mmio_return(vcpu, vcpu->run);
run               699 virt/kvm/arm/arm.c 	if (run->immediate_exit)
run               707 virt/kvm/arm/arm.c 	run->exit_reason = KVM_EXIT_UNKNOWN;
run               737 virt/kvm/arm/arm.c 			run->exit_reason = KVM_EXIT_INTR;
run               751 virt/kvm/arm/arm.c 				run->exit_reason = KVM_EXIT_INTR;
run               848 virt/kvm/arm/arm.c 		handle_exit_early(vcpu, run, ret);
run               852 virt/kvm/arm/arm.c 		ret = handle_exit(vcpu, run, ret);
run                83 virt/kvm/arm/mmio.c int kvm_handle_mmio_return(struct kvm_vcpu *vcpu, struct kvm_run *run)
run                95 virt/kvm/arm/mmio.c 	if (!run->mmio.is_write) {
run                96 virt/kvm/arm/mmio.c 		len = run->mmio.len;
run               100 virt/kvm/arm/mmio.c 		data = kvm_mmio_read_buf(run->mmio.data, len);
run               111 virt/kvm/arm/mmio.c 		trace_kvm_mmio(KVM_TRACE_MMIO_READ, len, run->mmio.phys_addr,
run               156 virt/kvm/arm/mmio.c int io_mem_abort(struct kvm_vcpu *vcpu, struct kvm_run *run,
run               200 virt/kvm/arm/mmio.c 	run->mmio.is_write	= is_write;
run               201 virt/kvm/arm/mmio.c 	run->mmio.phys_addr	= fault_ipa;
run               202 virt/kvm/arm/mmio.c 	run->mmio.len		= len;
run               208 virt/kvm/arm/mmio.c 			memcpy(run->mmio.data, data_buf, len);
run               210 virt/kvm/arm/mmio.c 		kvm_handle_mmio_return(vcpu, run);
run               215 virt/kvm/arm/mmio.c 		memcpy(run->mmio.data, data_buf, len);
run               217 virt/kvm/arm/mmio.c 	run->exit_reason	= KVM_EXIT_MMIO;
run              1908 virt/kvm/arm/mmu.c int kvm_handle_guest_abort(struct kvm_vcpu *vcpu, struct kvm_run *run)
run              1987 virt/kvm/arm/mmu.c 		ret = io_mem_abort(vcpu, run, fault_ipa);
run               392 virt/kvm/arm/pmu.c 	struct kvm_sync_regs *sregs = &vcpu->run->s.regs;
run               406 virt/kvm/arm/pmu.c 	struct kvm_sync_regs *regs = &vcpu->run->s.regs;
run               206 virt/kvm/arm/psci.c 	memset(&vcpu->run->system_event, 0, sizeof(vcpu->run->system_event));
run               207 virt/kvm/arm/psci.c 	vcpu->run->system_event.type = type;
run               208 virt/kvm/arm/psci.c 	vcpu->run->exit_reason = KVM_EXIT_SYSTEM_EVENT;
run               345 virt/kvm/kvm_main.c 	vcpu->run = page_address(page);
run               358 virt/kvm/kvm_main.c 	free_page((unsigned long)vcpu->run);
run               373 virt/kvm/kvm_main.c 	free_page((unsigned long)vcpu->run);
run              2717 virt/kvm/kvm_main.c 		page = virt_to_page(vcpu->run);
run              2919 virt/kvm/kvm_main.c 		r = kvm_arch_vcpu_ioctl_run(vcpu, vcpu->run);
run              2920 virt/kvm/kvm_main.c 		trace_kvm_userspace_exit(vcpu->run->exit_reason, r);