Lines Matching refs:vcpu
23 static int diag_release_pages(struct kvm_vcpu *vcpu) in diag_release_pages() argument
26 unsigned long prefix = kvm_s390_get_prefix(vcpu); in diag_release_pages()
28 start = vcpu->run->s.regs.gprs[(vcpu->arch.sie_block->ipa & 0xf0) >> 4]; in diag_release_pages()
29 end = vcpu->run->s.regs.gprs[vcpu->arch.sie_block->ipa & 0xf] + 4096; in diag_release_pages()
30 vcpu->stat.diagnose_10++; in diag_release_pages()
34 return kvm_s390_inject_program_int(vcpu, PGM_SPECIFICATION); in diag_release_pages()
36 VCPU_EVENT(vcpu, 5, "diag release pages %lX %lX", start, end); in diag_release_pages()
43 gmap_discard(vcpu->arch.gmap, start, end); in diag_release_pages()
51 gmap_discard(vcpu->arch.gmap, start, prefix); in diag_release_pages()
53 gmap_discard(vcpu->arch.gmap, 0, 4096); in diag_release_pages()
55 gmap_discard(vcpu->arch.gmap, 4096, 8192); in diag_release_pages()
56 gmap_discard(vcpu->arch.gmap, prefix + 2 * PAGE_SIZE, end); in diag_release_pages()
61 static int __diag_page_ref_service(struct kvm_vcpu *vcpu) in __diag_page_ref_service() argument
75 u16 rx = (vcpu->arch.sie_block->ipa & 0xf0) >> 4; in __diag_page_ref_service()
76 u16 ry = (vcpu->arch.sie_block->ipa & 0x0f); in __diag_page_ref_service()
78 VCPU_EVENT(vcpu, 3, "diag page reference parameter block at 0x%llx", in __diag_page_ref_service()
79 vcpu->run->s.regs.gprs[rx]); in __diag_page_ref_service()
80 vcpu->stat.diagnose_258++; in __diag_page_ref_service()
81 if (vcpu->run->s.regs.gprs[rx] & 7) in __diag_page_ref_service()
82 return kvm_s390_inject_program_int(vcpu, PGM_SPECIFICATION); in __diag_page_ref_service()
83 rc = read_guest(vcpu, vcpu->run->s.regs.gprs[rx], rx, &parm, sizeof(parm)); in __diag_page_ref_service()
85 return kvm_s390_inject_prog_cond(vcpu, rc); in __diag_page_ref_service()
87 return kvm_s390_inject_program_int(vcpu, PGM_SPECIFICATION); in __diag_page_ref_service()
91 VCPU_EVENT(vcpu, 3, "pageref token addr 0x%llx " in __diag_page_ref_service()
94 if (vcpu->arch.pfault_token != KVM_S390_PFAULT_TOKEN_INVALID) { in __diag_page_ref_service()
100 vcpu->run->s.regs.gprs[ry] = 8; in __diag_page_ref_service()
106 return kvm_s390_inject_program_int(vcpu, PGM_SPECIFICATION); in __diag_page_ref_service()
108 if (kvm_is_error_gpa(vcpu->kvm, parm.token_addr)) in __diag_page_ref_service()
109 return kvm_s390_inject_program_int(vcpu, PGM_ADDRESSING); in __diag_page_ref_service()
111 vcpu->arch.pfault_token = parm.token_addr; in __diag_page_ref_service()
112 vcpu->arch.pfault_select = parm.select_mask; in __diag_page_ref_service()
113 vcpu->arch.pfault_compare = parm.compare_mask; in __diag_page_ref_service()
114 vcpu->run->s.regs.gprs[ry] = 0; in __diag_page_ref_service()
123 VCPU_EVENT(vcpu, 3, "pageref cancel addr 0x%llx", parm.token_addr); in __diag_page_ref_service()
126 return kvm_s390_inject_program_int(vcpu, PGM_SPECIFICATION); in __diag_page_ref_service()
128 vcpu->run->s.regs.gprs[ry] = 0; in __diag_page_ref_service()
133 if (vcpu->arch.pfault_token == KVM_S390_PFAULT_TOKEN_INVALID) in __diag_page_ref_service()
134 vcpu->run->s.regs.gprs[ry] = 4; in __diag_page_ref_service()
136 vcpu->arch.pfault_token = KVM_S390_PFAULT_TOKEN_INVALID; in __diag_page_ref_service()
148 static int __diag_time_slice_end(struct kvm_vcpu *vcpu) in __diag_time_slice_end() argument
150 VCPU_EVENT(vcpu, 5, "%s", "diag time slice end"); in __diag_time_slice_end()
151 vcpu->stat.diagnose_44++; in __diag_time_slice_end()
152 kvm_vcpu_on_spin(vcpu); in __diag_time_slice_end()
156 static int __diag_time_slice_end_directed(struct kvm_vcpu *vcpu) in __diag_time_slice_end_directed() argument
158 struct kvm *kvm = vcpu->kvm; in __diag_time_slice_end_directed()
163 tid = vcpu->run->s.regs.gprs[(vcpu->arch.sie_block->ipa & 0xf0) >> 4]; in __diag_time_slice_end_directed()
164 vcpu->stat.diagnose_9c++; in __diag_time_slice_end_directed()
165 VCPU_EVENT(vcpu, 5, "diag time slice end directed to %d", tid); in __diag_time_slice_end_directed()
167 if (tid == vcpu->vcpu_id) in __diag_time_slice_end_directed()
179 static int __diag_ipl_functions(struct kvm_vcpu *vcpu) in __diag_ipl_functions() argument
181 unsigned int reg = vcpu->arch.sie_block->ipa & 0xf; in __diag_ipl_functions()
182 unsigned long subcode = vcpu->run->s.regs.gprs[reg] & 0xffff; in __diag_ipl_functions()
184 VCPU_EVENT(vcpu, 3, "diag ipl functions, subcode %lx", subcode); in __diag_ipl_functions()
185 vcpu->stat.diagnose_308++; in __diag_ipl_functions()
188 vcpu->run->s390_reset_flags = KVM_S390_RESET_CLEAR; in __diag_ipl_functions()
191 vcpu->run->s390_reset_flags = 0; in __diag_ipl_functions()
197 if (!kvm_s390_user_cpu_state_ctrl(vcpu->kvm)) in __diag_ipl_functions()
198 kvm_s390_vcpu_stop(vcpu); in __diag_ipl_functions()
199 vcpu->run->s390_reset_flags |= KVM_S390_RESET_SUBSYSTEM; in __diag_ipl_functions()
200 vcpu->run->s390_reset_flags |= KVM_S390_RESET_IPL; in __diag_ipl_functions()
201 vcpu->run->s390_reset_flags |= KVM_S390_RESET_CPU_INIT; in __diag_ipl_functions()
202 vcpu->run->exit_reason = KVM_EXIT_S390_RESET; in __diag_ipl_functions()
203 VCPU_EVENT(vcpu, 3, "requesting userspace resets %llx", in __diag_ipl_functions()
204 vcpu->run->s390_reset_flags); in __diag_ipl_functions()
205 trace_kvm_s390_request_resets(vcpu->run->s390_reset_flags); in __diag_ipl_functions()
209 static int __diag_virtio_hypercall(struct kvm_vcpu *vcpu) in __diag_virtio_hypercall() argument
213 vcpu->stat.diagnose_500++; in __diag_virtio_hypercall()
215 if (!vcpu->kvm->arch.css_support || in __diag_virtio_hypercall()
216 (vcpu->run->s.regs.gprs[1] != KVM_S390_VIRTIO_CCW_NOTIFY)) in __diag_virtio_hypercall()
225 ret = kvm_io_bus_write_cookie(vcpu, KVM_VIRTIO_CCW_NOTIFY_BUS, in __diag_virtio_hypercall()
226 vcpu->run->s.regs.gprs[2] & 0xffffffff, in __diag_virtio_hypercall()
227 8, &vcpu->run->s.regs.gprs[3], in __diag_virtio_hypercall()
228 vcpu->run->s.regs.gprs[4]); in __diag_virtio_hypercall()
235 vcpu->run->s.regs.gprs[2] = ret; in __diag_virtio_hypercall()
240 int kvm_s390_handle_diag(struct kvm_vcpu *vcpu) in kvm_s390_handle_diag() argument
242 int code = kvm_s390_get_base_disp_rs(vcpu, NULL) & 0xffff; in kvm_s390_handle_diag()
244 if (vcpu->arch.sie_block->gpsw.mask & PSW_MASK_PSTATE) in kvm_s390_handle_diag()
245 return kvm_s390_inject_program_int(vcpu, PGM_PRIVILEGED_OP); in kvm_s390_handle_diag()
247 trace_kvm_s390_handle_diag(vcpu, code); in kvm_s390_handle_diag()
250 return diag_release_pages(vcpu); in kvm_s390_handle_diag()
252 return __diag_time_slice_end(vcpu); in kvm_s390_handle_diag()
254 return __diag_time_slice_end_directed(vcpu); in kvm_s390_handle_diag()
256 return __diag_page_ref_service(vcpu); in kvm_s390_handle_diag()
258 return __diag_ipl_functions(vcpu); in kvm_s390_handle_diag()
260 return __diag_virtio_hypercall(vcpu); in kvm_s390_handle_diag()