msr_index 212 arch/x86/include/asm/kvm_emulate.h int (*set_msr)(struct x86_emulate_ctxt *ctxt, u32 msr_index, u64 data); msr_index 213 arch/x86/include/asm/kvm_emulate.h int (*get_msr)(struct x86_emulate_ctxt *ctxt, u32 msr_index, u64 *pdata); msr_index 1087 arch/x86/kvm/vmx/nested.c vmx_restore_control_msr(struct vcpu_vmx *vmx, u32 msr_index, u64 data) msr_index 1092 arch/x86/kvm/vmx/nested.c switch (msr_index) { msr_index 1185 arch/x86/kvm/vmx/nested.c static int vmx_restore_fixed0_msr(struct vcpu_vmx *vmx, u32 msr_index, u64 data) msr_index 1189 arch/x86/kvm/vmx/nested.c switch (msr_index) { msr_index 1216 arch/x86/kvm/vmx/nested.c int vmx_set_vmx_msr(struct kvm_vcpu *vcpu, u32 msr_index, u64 data) msr_index 1227 arch/x86/kvm/vmx/nested.c switch (msr_index) { msr_index 1249 arch/x86/kvm/vmx/nested.c return vmx_restore_control_msr(vmx, msr_index, data); msr_index 1254 arch/x86/kvm/vmx/nested.c return vmx_restore_fixed0_msr(vmx, msr_index, data); msr_index 1281 arch/x86/kvm/vmx/nested.c int vmx_get_vmx_msr(struct nested_vmx_msrs *msrs, u32 msr_index, u64 *pdata) msr_index 1283 arch/x86/kvm/vmx/nested.c switch (msr_index) { msr_index 1292 arch/x86/kvm/vmx/nested.c if (msr_index == MSR_IA32_VMX_PINBASED_CTLS) msr_index 1300 arch/x86/kvm/vmx/nested.c if (msr_index == MSR_IA32_VMX_PROCBASED_CTLS) msr_index 1308 arch/x86/kvm/vmx/nested.c if (msr_index == MSR_IA32_VMX_EXIT_CTLS) msr_index 1316 arch/x86/kvm/vmx/nested.c if (msr_index == MSR_IA32_VMX_ENTRY_CTLS) msr_index 5182 arch/x86/kvm/vmx/nested.c u32 msr_index = kvm_rcx_read(vcpu); msr_index 5196 arch/x86/kvm/vmx/nested.c if (msr_index >= 0xc0000000) { msr_index 5197 arch/x86/kvm/vmx/nested.c msr_index -= 0xc0000000; msr_index 5202 arch/x86/kvm/vmx/nested.c if (msr_index < 1024*8) { msr_index 5204 arch/x86/kvm/vmx/nested.c if (kvm_vcpu_read_guest(vcpu, bitmap + msr_index/8, &b, 1)) msr_index 5206 arch/x86/kvm/vmx/nested.c return 1 & (b >> (msr_index & 7)); msr_index 31 arch/x86/kvm/vmx/nested.h int vmx_set_vmx_msr(struct kvm_vcpu *vcpu, u32 msr_index, u64 data); msr_index 32 arch/x86/kvm/vmx/nested.h int vmx_get_vmx_msr(struct nested_vmx_msrs *msrs, u32 msr_index, u64 *pdata); msr_index 1895 arch/x86/kvm/vmx/vmx.c u32 msr_index = msr_info->index; msr_index 1899 arch/x86/kvm/vmx/vmx.c switch (msr_index) { msr_index 2056 arch/x86/kvm/vmx/vmx.c return vmx_set_vmx_msr(vcpu, msr_index, data); msr_index 2144 arch/x86/kvm/vmx/vmx.c msr = find_msr_entry(vmx, msr_index); msr_index 6186 arch/x86/kvm/x86.c u32 msr_index, u64 *pdata) msr_index 6188 arch/x86/kvm/x86.c return kvm_get_msr(emul_to_vcpu(ctxt), msr_index, pdata); msr_index 6192 arch/x86/kvm/x86.c u32 msr_index, u64 data) msr_index 6194 arch/x86/kvm/x86.c return kvm_set_msr(emul_to_vcpu(ctxt), msr_index, data); msr_index 143 arch/x86/xen/pmu.c static int is_intel_pmu_msr(u32 msr_index, int *type, int *index) msr_index 147 arch/x86/xen/pmu.c switch (msr_index) { msr_index 162 arch/x86/xen/pmu.c if ((msr_index >= MSR_CORE_PERF_FIXED_CTR0) && msr_index 163 arch/x86/xen/pmu.c (msr_index < MSR_CORE_PERF_FIXED_CTR0 + msr_index 165 arch/x86/xen/pmu.c *index = msr_index - MSR_CORE_PERF_FIXED_CTR0; msr_index 170 arch/x86/xen/pmu.c if ((msr_index >= MSR_P6_EVNTSEL0) && msr_index 171 arch/x86/xen/pmu.c (msr_index < MSR_P6_EVNTSEL0 + intel_num_arch_counters)) { msr_index 172 arch/x86/xen/pmu.c *index = msr_index - MSR_P6_EVNTSEL0; msr_index 177 arch/x86/xen/pmu.c msr_index_pmc = msr_index & MSR_PMC_ALIAS_MASK; msr_index 210 drivers/thermal/intel/intel_powerclamp.c int msr_index; msr_index 215 drivers/thermal/intel/intel_powerclamp.c .msr_index = MSR_PKG_C##id##_RESIDENCY, \ msr_index 236 drivers/thermal/intel/intel_powerclamp.c while (info->msr_index) { msr_index 237 drivers/thermal/intel/intel_powerclamp.c if (!rdmsrl_safe(info->msr_index, &val)) msr_index 251 drivers/thermal/intel/intel_powerclamp.c while (info->msr_index) { msr_index 253 drivers/thermal/intel/intel_powerclamp.c if (!rdmsrl_safe(info->msr_index, &val)) msr_index 324 tools/testing/selftests/kvm/include/x86_64/processor.h uint64_t vcpu_get_msr(struct kvm_vm *vm, uint32_t vcpuid, uint64_t msr_index); msr_index 325 tools/testing/selftests/kvm/include/x86_64/processor.h void vcpu_set_msr(struct kvm_vm *vm, uint32_t vcpuid, uint64_t msr_index, msr_index 853 tools/testing/selftests/kvm/lib/x86_64/processor.c uint64_t vcpu_get_msr(struct kvm_vm *vm, uint32_t vcpuid, uint64_t msr_index) msr_index 864 tools/testing/selftests/kvm/lib/x86_64/processor.c buffer.entry.index = msr_index; msr_index 886 tools/testing/selftests/kvm/lib/x86_64/processor.c void vcpu_set_msr(struct kvm_vm *vm, uint32_t vcpuid, uint64_t msr_index, msr_index 899 tools/testing/selftests/kvm/lib/x86_64/processor.c buffer.entry.index = msr_index;