Lines Matching refs:sie_block

65 	u64 *cr9 = &vcpu->arch.sie_block->gcr[9];  in enable_all_hw_bp()
66 u64 *cr10 = &vcpu->arch.sie_block->gcr[10]; in enable_all_hw_bp()
67 u64 *cr11 = &vcpu->arch.sie_block->gcr[11]; in enable_all_hw_bp()
105 u64 *cr9 = &vcpu->arch.sie_block->gcr[9]; in enable_all_hw_wp()
106 u64 *cr10 = &vcpu->arch.sie_block->gcr[10]; in enable_all_hw_wp()
107 u64 *cr11 = &vcpu->arch.sie_block->gcr[11]; in enable_all_hw_wp()
135 vcpu->arch.guestdbg.cr0 = vcpu->arch.sie_block->gcr[0]; in kvm_s390_backup_guest_per_regs()
136 vcpu->arch.guestdbg.cr9 = vcpu->arch.sie_block->gcr[9]; in kvm_s390_backup_guest_per_regs()
137 vcpu->arch.guestdbg.cr10 = vcpu->arch.sie_block->gcr[10]; in kvm_s390_backup_guest_per_regs()
138 vcpu->arch.guestdbg.cr11 = vcpu->arch.sie_block->gcr[11]; in kvm_s390_backup_guest_per_regs()
143 vcpu->arch.sie_block->gcr[0] = vcpu->arch.guestdbg.cr0; in kvm_s390_restore_guest_per_regs()
144 vcpu->arch.sie_block->gcr[9] = vcpu->arch.guestdbg.cr9; in kvm_s390_restore_guest_per_regs()
145 vcpu->arch.sie_block->gcr[10] = vcpu->arch.guestdbg.cr10; in kvm_s390_restore_guest_per_regs()
146 vcpu->arch.sie_block->gcr[11] = vcpu->arch.guestdbg.cr11; in kvm_s390_restore_guest_per_regs()
159 vcpu->arch.sie_block->gcr[0] &= ~0x800ul; in kvm_s390_patch_guest_per_regs()
160 vcpu->arch.sie_block->gcr[9] |= PER_EVENT_IFETCH; in kvm_s390_patch_guest_per_regs()
161 vcpu->arch.sie_block->gcr[10] = 0; in kvm_s390_patch_guest_per_regs()
162 vcpu->arch.sie_block->gcr[11] = PSW_ADDR_INSN; in kvm_s390_patch_guest_per_regs()
171 if (vcpu->arch.sie_block->gcr[9] & PER_EVENT_NULLIFICATION) in kvm_s390_patch_guest_per_regs()
172 vcpu->arch.sie_block->gcr[9] &= ~PER_EVENT_NULLIFICATION; in kvm_s390_patch_guest_per_regs()
392 u32 perc = (vcpu->arch.sie_block->perc << 24); in debug_exit_required()
396 unsigned long addr = vcpu->arch.sie_block->gpsw.addr; in debug_exit_required()
397 unsigned long peraddr = vcpu->arch.sie_block->peraddr; in debug_exit_required()
440 (vcpu->arch.sie_block->gpsw.mask & PSW_MASK_PER)
444 u32 perc = vcpu->arch.sie_block->perc << 24; in filter_guest_per_event()
445 u64 peraddr = vcpu->arch.sie_block->peraddr; in filter_guest_per_event()
446 u64 addr = vcpu->arch.sie_block->gpsw.addr; in filter_guest_per_event()
447 u64 cr9 = vcpu->arch.sie_block->gcr[9]; in filter_guest_per_event()
448 u64 cr10 = vcpu->arch.sie_block->gcr[10]; in filter_guest_per_event()
449 u64 cr11 = vcpu->arch.sie_block->gcr[11]; in filter_guest_per_event()
470 vcpu->arch.sie_block->perc = guest_perc >> 24; in filter_guest_per_event()
473 vcpu->arch.sie_block->iprcc &= ~PGM_PER; in filter_guest_per_event()
476 #define pssec(vcpu) (vcpu->arch.sie_block->gcr[1] & _ASCE_SPACE_SWITCH)
477 #define hssec(vcpu) (vcpu->arch.sie_block->gcr[13] & _ASCE_SPACE_SWITCH)
478 #define old_ssec(vcpu) ((vcpu->arch.sie_block->tecmc >> 31) & 0x1)
479 #define old_as_is_home(vcpu) !(vcpu->arch.sie_block->tecmc & 0xffff)
496 if (vcpu->arch.sie_block->iprcc == PGM_SPACE_SWITCH) { in kvm_s390_handle_per_event()
497 vcpu->arch.sie_block->iprcc = 0; in kvm_s390_handle_per_event()
498 new_as = psw_bits(vcpu->arch.sie_block->gpsw).as; in kvm_s390_handle_per_event()
507 vcpu->arch.sie_block->iprcc = PGM_SPACE_SWITCH; in kvm_s390_handle_per_event()
515 vcpu->arch.sie_block->iprcc = PGM_SPACE_SWITCH; in kvm_s390_handle_per_event()