msr_index         212 arch/x86/include/asm/kvm_emulate.h 	int (*set_msr)(struct x86_emulate_ctxt *ctxt, u32 msr_index, u64 data);
msr_index         213 arch/x86/include/asm/kvm_emulate.h 	int (*get_msr)(struct x86_emulate_ctxt *ctxt, u32 msr_index, u64 *pdata);
msr_index        1087 arch/x86/kvm/vmx/nested.c vmx_restore_control_msr(struct vcpu_vmx *vmx, u32 msr_index, u64 data)
msr_index        1092 arch/x86/kvm/vmx/nested.c 	switch (msr_index) {
msr_index        1185 arch/x86/kvm/vmx/nested.c static int vmx_restore_fixed0_msr(struct vcpu_vmx *vmx, u32 msr_index, u64 data)
msr_index        1189 arch/x86/kvm/vmx/nested.c 	switch (msr_index) {
msr_index        1216 arch/x86/kvm/vmx/nested.c int vmx_set_vmx_msr(struct kvm_vcpu *vcpu, u32 msr_index, u64 data)
msr_index        1227 arch/x86/kvm/vmx/nested.c 	switch (msr_index) {
msr_index        1249 arch/x86/kvm/vmx/nested.c 		return vmx_restore_control_msr(vmx, msr_index, data);
msr_index        1254 arch/x86/kvm/vmx/nested.c 		return vmx_restore_fixed0_msr(vmx, msr_index, data);
msr_index        1281 arch/x86/kvm/vmx/nested.c int vmx_get_vmx_msr(struct nested_vmx_msrs *msrs, u32 msr_index, u64 *pdata)
msr_index        1283 arch/x86/kvm/vmx/nested.c 	switch (msr_index) {
msr_index        1292 arch/x86/kvm/vmx/nested.c 		if (msr_index == MSR_IA32_VMX_PINBASED_CTLS)
msr_index        1300 arch/x86/kvm/vmx/nested.c 		if (msr_index == MSR_IA32_VMX_PROCBASED_CTLS)
msr_index        1308 arch/x86/kvm/vmx/nested.c 		if (msr_index == MSR_IA32_VMX_EXIT_CTLS)
msr_index        1316 arch/x86/kvm/vmx/nested.c 		if (msr_index == MSR_IA32_VMX_ENTRY_CTLS)
msr_index        5182 arch/x86/kvm/vmx/nested.c 	u32 msr_index = kvm_rcx_read(vcpu);
msr_index        5196 arch/x86/kvm/vmx/nested.c 	if (msr_index >= 0xc0000000) {
msr_index        5197 arch/x86/kvm/vmx/nested.c 		msr_index -= 0xc0000000;
msr_index        5202 arch/x86/kvm/vmx/nested.c 	if (msr_index < 1024*8) {
msr_index        5204 arch/x86/kvm/vmx/nested.c 		if (kvm_vcpu_read_guest(vcpu, bitmap + msr_index/8, &b, 1))
msr_index        5206 arch/x86/kvm/vmx/nested.c 		return 1 & (b >> (msr_index & 7));
msr_index          31 arch/x86/kvm/vmx/nested.h int vmx_set_vmx_msr(struct kvm_vcpu *vcpu, u32 msr_index, u64 data);
msr_index          32 arch/x86/kvm/vmx/nested.h int vmx_get_vmx_msr(struct nested_vmx_msrs *msrs, u32 msr_index, u64 *pdata);
msr_index        1895 arch/x86/kvm/vmx/vmx.c 	u32 msr_index = msr_info->index;
msr_index        1899 arch/x86/kvm/vmx/vmx.c 	switch (msr_index) {
msr_index        2056 arch/x86/kvm/vmx/vmx.c 		return vmx_set_vmx_msr(vcpu, msr_index, data);
msr_index        2144 arch/x86/kvm/vmx/vmx.c 		msr = find_msr_entry(vmx, msr_index);
msr_index        6186 arch/x86/kvm/x86.c 			    u32 msr_index, u64 *pdata)
msr_index        6188 arch/x86/kvm/x86.c 	return kvm_get_msr(emul_to_vcpu(ctxt), msr_index, pdata);
msr_index        6192 arch/x86/kvm/x86.c 			    u32 msr_index, u64 data)
msr_index        6194 arch/x86/kvm/x86.c 	return kvm_set_msr(emul_to_vcpu(ctxt), msr_index, data);
msr_index         143 arch/x86/xen/pmu.c static int is_intel_pmu_msr(u32 msr_index, int *type, int *index)
msr_index         147 arch/x86/xen/pmu.c 	switch (msr_index) {
msr_index         162 arch/x86/xen/pmu.c 		if ((msr_index >= MSR_CORE_PERF_FIXED_CTR0) &&
msr_index         163 arch/x86/xen/pmu.c 		    (msr_index < MSR_CORE_PERF_FIXED_CTR0 +
msr_index         165 arch/x86/xen/pmu.c 			*index = msr_index - MSR_CORE_PERF_FIXED_CTR0;
msr_index         170 arch/x86/xen/pmu.c 		if ((msr_index >= MSR_P6_EVNTSEL0) &&
msr_index         171 arch/x86/xen/pmu.c 		    (msr_index < MSR_P6_EVNTSEL0 +  intel_num_arch_counters)) {
msr_index         172 arch/x86/xen/pmu.c 			*index = msr_index - MSR_P6_EVNTSEL0;
msr_index         177 arch/x86/xen/pmu.c 		msr_index_pmc = msr_index & MSR_PMC_ALIAS_MASK;
msr_index         210 drivers/thermal/intel/intel_powerclamp.c 	int msr_index;
msr_index         215 drivers/thermal/intel/intel_powerclamp.c 		.msr_index = MSR_PKG_C##id##_RESIDENCY, \
msr_index         236 drivers/thermal/intel/intel_powerclamp.c 	while (info->msr_index) {
msr_index         237 drivers/thermal/intel/intel_powerclamp.c 		if (!rdmsrl_safe(info->msr_index, &val))
msr_index         251 drivers/thermal/intel/intel_powerclamp.c 	while (info->msr_index) {
msr_index         253 drivers/thermal/intel/intel_powerclamp.c 			if (!rdmsrl_safe(info->msr_index, &val))
msr_index         324 tools/testing/selftests/kvm/include/x86_64/processor.h uint64_t vcpu_get_msr(struct kvm_vm *vm, uint32_t vcpuid, uint64_t msr_index);
msr_index         325 tools/testing/selftests/kvm/include/x86_64/processor.h void vcpu_set_msr(struct kvm_vm *vm, uint32_t vcpuid, uint64_t msr_index,
msr_index         853 tools/testing/selftests/kvm/lib/x86_64/processor.c uint64_t vcpu_get_msr(struct kvm_vm *vm, uint32_t vcpuid, uint64_t msr_index)
msr_index         864 tools/testing/selftests/kvm/lib/x86_64/processor.c 	buffer.entry.index = msr_index;
msr_index         886 tools/testing/selftests/kvm/lib/x86_64/processor.c void vcpu_set_msr(struct kvm_vm *vm, uint32_t vcpuid, uint64_t msr_index,
msr_index         899 tools/testing/selftests/kvm/lib/x86_64/processor.c 	buffer.entry.index = msr_index;