mdcr_el2 251 arch/arm64/include/asm/kvm_host.h u32 mdcr_el2; mdcr_el2 24 arch/arm64/kvm/debug.c static DEFINE_PER_CPU(u32, mdcr_el2); mdcr_el2 68 arch/arm64/kvm/debug.c __this_cpu_write(mdcr_el2, kvm_call_hyp_ret(__kvm_get_mdcr_el2)); mdcr_el2 104 arch/arm64/kvm/debug.c unsigned long mdscr, orig_mdcr_el2 = vcpu->arch.mdcr_el2; mdcr_el2 112 arch/arm64/kvm/debug.c vcpu->arch.mdcr_el2 = __this_cpu_read(mdcr_el2) & MDCR_EL2_HPMN_MASK; mdcr_el2 113 arch/arm64/kvm/debug.c vcpu->arch.mdcr_el2 |= (MDCR_EL2_TPM | mdcr_el2 122 arch/arm64/kvm/debug.c vcpu->arch.mdcr_el2 |= MDCR_EL2_TDE; mdcr_el2 194 arch/arm64/kvm/debug.c vcpu->arch.mdcr_el2 |= MDCR_EL2_TDA; mdcr_el2 201 arch/arm64/kvm/debug.c if (has_vhe() && orig_mdcr_el2 != vcpu->arch.mdcr_el2) mdcr_el2 202 arch/arm64/kvm/debug.c write_sysreg(vcpu->arch.mdcr_el2, mdcr_el2); mdcr_el2 204 arch/arm64/kvm/debug.c trace_kvm_arm_set_dreg32("MDCR_EL2", vcpu->arch.mdcr_el2); mdcr_el2 223 arch/arm64/kvm/hyp/debug-sr.c return read_sysreg(mdcr_el2); mdcr_el2 85 arch/arm64/kvm/hyp/switch.c write_sysreg(vcpu->arch.mdcr_el2, mdcr_el2); mdcr_el2 168 arch/arm64/kvm/hyp/switch.c u64 mdcr_el2 = read_sysreg(mdcr_el2); mdcr_el2 172 arch/arm64/kvm/hyp/switch.c mdcr_el2 &= MDCR_EL2_HPMN_MASK; mdcr_el2 173 arch/arm64/kvm/hyp/switch.c mdcr_el2 |= MDCR_EL2_E2PB_MASK << MDCR_EL2_E2PB_SHIFT; mdcr_el2 175 arch/arm64/kvm/hyp/switch.c write_sysreg(mdcr_el2, mdcr_el2); mdcr_el2 206 arch/arm64/kvm/hyp/switch.c u64 mdcr_el2 = read_sysreg(mdcr_el2); mdcr_el2 208 arch/arm64/kvm/hyp/switch.c mdcr_el2 &= MDCR_EL2_HPMN_MASK | mdcr_el2 212 arch/arm64/kvm/hyp/switch.c write_sysreg(mdcr_el2, mdcr_el2);