mdcr_el2          251 arch/arm64/include/asm/kvm_host.h 	u32 mdcr_el2;
mdcr_el2           24 arch/arm64/kvm/debug.c static DEFINE_PER_CPU(u32, mdcr_el2);
mdcr_el2           68 arch/arm64/kvm/debug.c 	__this_cpu_write(mdcr_el2, kvm_call_hyp_ret(__kvm_get_mdcr_el2));
mdcr_el2          104 arch/arm64/kvm/debug.c 	unsigned long mdscr, orig_mdcr_el2 = vcpu->arch.mdcr_el2;
mdcr_el2          112 arch/arm64/kvm/debug.c 	vcpu->arch.mdcr_el2 = __this_cpu_read(mdcr_el2) & MDCR_EL2_HPMN_MASK;
mdcr_el2          113 arch/arm64/kvm/debug.c 	vcpu->arch.mdcr_el2 |= (MDCR_EL2_TPM |
mdcr_el2          122 arch/arm64/kvm/debug.c 		vcpu->arch.mdcr_el2 |= MDCR_EL2_TDE;
mdcr_el2          194 arch/arm64/kvm/debug.c 		vcpu->arch.mdcr_el2 |= MDCR_EL2_TDA;
mdcr_el2          201 arch/arm64/kvm/debug.c 	if (has_vhe() && orig_mdcr_el2 != vcpu->arch.mdcr_el2)
mdcr_el2          202 arch/arm64/kvm/debug.c 		write_sysreg(vcpu->arch.mdcr_el2, mdcr_el2);
mdcr_el2          204 arch/arm64/kvm/debug.c 	trace_kvm_arm_set_dreg32("MDCR_EL2", vcpu->arch.mdcr_el2);
mdcr_el2          223 arch/arm64/kvm/hyp/debug-sr.c 	return read_sysreg(mdcr_el2);
mdcr_el2           85 arch/arm64/kvm/hyp/switch.c 	write_sysreg(vcpu->arch.mdcr_el2, mdcr_el2);
mdcr_el2          168 arch/arm64/kvm/hyp/switch.c 	u64 mdcr_el2 = read_sysreg(mdcr_el2);
mdcr_el2          172 arch/arm64/kvm/hyp/switch.c 	mdcr_el2 &= MDCR_EL2_HPMN_MASK;
mdcr_el2          173 arch/arm64/kvm/hyp/switch.c 	mdcr_el2 |= MDCR_EL2_E2PB_MASK << MDCR_EL2_E2PB_SHIFT;
mdcr_el2          175 arch/arm64/kvm/hyp/switch.c 	write_sysreg(mdcr_el2, mdcr_el2);
mdcr_el2          206 arch/arm64/kvm/hyp/switch.c 	u64 mdcr_el2 = read_sysreg(mdcr_el2);
mdcr_el2          208 arch/arm64/kvm/hyp/switch.c 	mdcr_el2 &= MDCR_EL2_HPMN_MASK |
mdcr_el2          212 arch/arm64/kvm/hyp/switch.c 	write_sysreg(mdcr_el2, mdcr_el2);