Searched refs:msr (Results 1 - 200 of 364) sorted by relevance

12

/linux-4.1.27/arch/x86/lib/
H A Dmsr-reg-export.c2 #include <asm/msr.h>
H A Dmsr.c3 #include <asm/msr.h>
5 struct msr *msrs_alloc(void) msrs_alloc()
7 struct msr *msrs = NULL; msrs_alloc()
9 msrs = alloc_percpu(struct msr); msrs_alloc()
19 void msrs_free(struct msr *msrs) msrs_free()
28 * @msr: MSR to read
35 int msr_read(u32 msr, struct msr *m) msr_read() argument
40 err = rdmsrl_safe(msr, &val); msr_read()
50 * @msr: MSR to write
53 int msr_write(u32 msr, struct msr *m) msr_write() argument
55 return wrmsrl_safe(msr, m->q); msr_write()
58 static inline int __flip_bit(u32 msr, u8 bit, bool set) __flip_bit() argument
60 struct msr m, m1; __flip_bit()
66 err = msr_read(msr, &m); __flip_bit()
79 err = msr_write(msr, &m1); __flip_bit()
87 * Set @bit in a MSR @msr.
94 int msr_set_bit(u32 msr, u8 bit) msr_set_bit() argument
96 return __flip_bit(msr, bit, true); msr_set_bit()
100 * Clear @bit in a MSR @msr.
107 int msr_clear_bit(u32 msr, u8 bit) msr_clear_bit() argument
109 return __flip_bit(msr, bit, false); msr_clear_bit()
H A DMakefile17 obj-$(CONFIG_SMP) += msr-smp.o cache-smp.o
26 obj-y += msr.o msr-reg.o msr-reg-export.o
H A Dmsr-smp.c4 #include <asm/msr.h>
9 struct msr *reg; __rdmsr_on_cpu()
23 struct msr *reg; __wrmsr_on_cpu()
98 struct msr *msrs, __rwmsr_on_cpus()
125 void rdmsr_on_cpus(const struct cpumask *mask, u32 msr_no, struct msr *msrs) rdmsr_on_cpus()
139 void wrmsr_on_cpus(const struct cpumask *mask, u32 msr_no, struct msr *msrs) wrmsr_on_cpus()
H A Dmsr-reg.S5 #include <asm/msr.h>
/linux-4.1.27/arch/x86/kernel/cpu/
H A Dperfctr-watchdog.c43 /* converts an msr to an appropriate reservation bit */ nmi_perfctr_msr_to_bit()
44 static inline unsigned int nmi_perfctr_msr_to_bit(unsigned int msr) nmi_perfctr_msr_to_bit() argument
49 if (msr >= MSR_F15H_PERF_CTR) nmi_perfctr_msr_to_bit()
50 return (msr - MSR_F15H_PERF_CTR) >> 1; nmi_perfctr_msr_to_bit()
51 return msr - MSR_K7_PERFCTR0; nmi_perfctr_msr_to_bit()
54 return msr - MSR_ARCH_PERFMON_PERFCTR0; nmi_perfctr_msr_to_bit()
58 return msr - MSR_P6_PERFCTR0; nmi_perfctr_msr_to_bit()
60 return msr - MSR_KNC_PERFCTR0; nmi_perfctr_msr_to_bit()
62 return msr - MSR_P4_BPU_PERFCTR0; nmi_perfctr_msr_to_bit()
69 * converts an msr to an appropriate reservation bit
72 static inline unsigned int nmi_evntsel_msr_to_bit(unsigned int msr) nmi_evntsel_msr_to_bit() argument
77 if (msr >= MSR_F15H_PERF_CTL) nmi_evntsel_msr_to_bit()
78 return (msr - MSR_F15H_PERF_CTL) >> 1; nmi_evntsel_msr_to_bit()
79 return msr - MSR_K7_EVNTSEL0; nmi_evntsel_msr_to_bit()
82 return msr - MSR_ARCH_PERFMON_EVENTSEL0; nmi_evntsel_msr_to_bit()
86 return msr - MSR_P6_EVNTSEL0; nmi_evntsel_msr_to_bit()
88 return msr - MSR_KNC_EVNTSEL0; nmi_evntsel_msr_to_bit()
90 return msr - MSR_P4_BSU_ESCR0; nmi_evntsel_msr_to_bit()
106 int reserve_perfctr_nmi(unsigned int msr) reserve_perfctr_nmi() argument
110 counter = nmi_perfctr_msr_to_bit(msr); reserve_perfctr_nmi()
121 void release_perfctr_nmi(unsigned int msr) release_perfctr_nmi() argument
125 counter = nmi_perfctr_msr_to_bit(msr); release_perfctr_nmi()
134 int reserve_evntsel_nmi(unsigned int msr) reserve_evntsel_nmi() argument
138 counter = nmi_evntsel_msr_to_bit(msr); reserve_evntsel_nmi()
149 void release_evntsel_nmi(unsigned int msr) release_evntsel_nmi() argument
153 counter = nmi_evntsel_msr_to_bit(msr); release_evntsel_nmi()
H A Dperf_event_intel_uncore_snbep.c100 .msr = SNBEP_C0_MSR_PMON_BOX_FILTER, \
325 unsigned msr; snbep_uncore_msr_disable_box() local
327 msr = uncore_msr_box_ctl(box); snbep_uncore_msr_disable_box()
328 if (msr) { snbep_uncore_msr_disable_box()
329 rdmsrl(msr, config); snbep_uncore_msr_disable_box()
331 wrmsrl(msr, config); snbep_uncore_msr_disable_box()
338 unsigned msr; snbep_uncore_msr_enable_box() local
340 msr = uncore_msr_box_ctl(box); snbep_uncore_msr_enable_box()
341 if (msr) { snbep_uncore_msr_enable_box()
342 rdmsrl(msr, config); snbep_uncore_msr_enable_box()
344 wrmsrl(msr, config); snbep_uncore_msr_enable_box()
369 unsigned msr = uncore_msr_box_ctl(box); snbep_uncore_msr_init_box() local
371 if (msr) snbep_uncore_msr_init_box()
372 wrmsrl(msr, SNBEP_PMON_BOX_CTL_INT); snbep_uncore_msr_init_box()
731 for (er = snbep_uncore_cbox_extra_regs; er->msr; er++) { snbep_cbox_hw_config()
1154 unsigned msr = uncore_msr_box_ctl(box); ivbep_uncore_msr_init_box() local
1155 if (msr) ivbep_uncore_msr_init_box()
1156 wrmsrl(msr, IVBEP_PMON_BOX_CTL_INT); ivbep_uncore_msr_init_box()
1385 for (er = ivbep_uncore_cbox_extra_regs; er->msr; er++) { ivbep_cbox_hw_config()
1871 for (er = hswep_uncore_cbox_extra_regs; er->msr; er++) { hswep_cbox_hw_config()
1934 unsigned msr = uncore_msr_box_ctl(box); hswep_uncore_sbox_msr_init_box() local
1936 if (msr) { hswep_uncore_sbox_msr_init_box()
1943 wrmsrl(msr, flags); hswep_uncore_sbox_msr_init_box()
H A Dbugs.c16 #include <asm/msr.h>
H A Dtransmeta.c4 #include <asm/msr.h>
H A Dperf_event_intel_uncore_nhmex.c206 unsigned msr = uncore_msr_box_ctl(box); nhmex_uncore_msr_disable_box() local
209 if (msr) { nhmex_uncore_msr_disable_box()
210 rdmsrl(msr, config); nhmex_uncore_msr_disable_box()
215 wrmsrl(msr, config); nhmex_uncore_msr_disable_box()
221 unsigned msr = uncore_msr_box_ctl(box); nhmex_uncore_msr_enable_box() local
224 if (msr) { nhmex_uncore_msr_enable_box()
225 rdmsrl(msr, config); nhmex_uncore_msr_enable_box()
230 wrmsrl(msr, config); nhmex_uncore_msr_enable_box()
301 /* msr offset for each instance of cbox */
764 unsigned msr; nhmex_mbox_hw_config() local
771 for (er = nhmex_uncore_mbox_extra_regs; er->msr; er++) { nhmex_mbox_hw_config()
777 msr = er->msr + type->msr_offset * box->pmu->pmu_idx; nhmex_mbox_hw_config()
778 if (WARN_ON_ONCE(msr >= 0xffff || er->idx >= 0xff)) nhmex_mbox_hw_config()
790 reg1->reg |= msr << (reg_idx * 16); nhmex_mbox_hw_config()
H A Dperf_event_amd_ibs.c46 unsigned int msr; member in struct:perf_ibs
284 hwc->config_base = perf_ibs->msr; perf_ibs_init()
479 .msr = MSR_AMD64_IBSFETCHCTL,
503 .msr = MSR_AMD64_IBSOPCTL,
526 unsigned int msr; perf_ibs_handle_irq() local
539 msr = hwc->config_base; perf_ibs_handle_irq()
541 rdmsrl(msr, *buf); perf_ibs_handle_irq()
562 rdmsrl(msr + offset, *buf++); perf_ibs_handle_irq()
795 * the offset in the IBS_CTL per-node msr. The per-core APIC setup of
H A Dperf_event.h19 #define wrmsrl(msr, val) \
21 unsigned int _msr = (msr); \
407 unsigned int msr; member in struct:extra_reg
416 .msr = (ms), \
423 #define INTEL_EVENT_EXTRA_REG(event, msr, vm, idx) \
424 EVENT_EXTRA_REG(event, msr, ARCH_PERFMON_EVENTSEL_EVENT, vm, idx)
426 #define INTEL_UEVENT_EXTRA_REG(event, msr, vm, idx) \
427 EVENT_EXTRA_REG(event, msr, ARCH_PERFMON_EVENTSEL_EVENT | \
H A Dperf_event_intel_rapl.c334 int bit, msr, ret = 0; rapl_pmu_event_init() local
350 msr = MSR_PP0_ENERGY_STATUS; rapl_pmu_event_init()
354 msr = MSR_PKG_ENERGY_STATUS; rapl_pmu_event_init()
358 msr = MSR_DRAM_ENERGY_STATUS; rapl_pmu_event_init()
362 msr = MSR_PP1_ENERGY_STATUS; rapl_pmu_event_init()
382 event->hw.event_base = msr; rapl_pmu_event_init()
H A Damd.c22 static inline int rdmsrl_amd_safe(unsigned msr, unsigned long long *p) rdmsrl_amd_safe() argument
30 gprs[1] = msr; rdmsrl_amd_safe()
40 static inline int wrmsrl_amd_safe(unsigned msr, unsigned long long val) wrmsrl_amd_safe() argument
48 gprs[1] = msr; wrmsrl_amd_safe()
577 * bit 6 of msr C001_0015 init_amd_k8()
H A Dcentaur.c7 #include <asm/msr.h>
/linux-4.1.27/arch/arm64/include/asm/
H A Dirqflags.h31 "msr daifset, #2" arch_local_irq_save()
41 "msr daifclr, #2 // arch_local_irq_enable" arch_local_irq_enable()
50 "msr daifset, #2 // arch_local_irq_disable" arch_local_irq_disable()
56 #define local_fiq_enable() asm("msr daifclr, #1" : : : "memory")
57 #define local_fiq_disable() asm("msr daifset, #1" : : : "memory")
59 #define local_async_enable() asm("msr daifclr, #4" : : : "memory")
60 #define local_async_disable() asm("msr daifset, #4" : : : "memory")
82 "msr daif, %0 // arch_local_irq_restore" arch_local_irq_restore()
101 "msr daifset, #8" \
109 "msr daif, %0 // local_dbg_restore\n" \
113 #define local_dbg_enable() asm("msr daifclr, #8" : : : "memory")
114 #define local_dbg_disable() asm("msr daifset, #8" : : : "memory")
H A Dassembler.h45 msr daifset, #2
49 msr daifclr, #2
61 msr daif, \olddaif
68 msr daifset, #8
72 msr daifclr, #8
79 msr mdscr_el1, \tmp
89 msr mdscr_el1, \tmp
99 msr daifclr, #(8 | 2)
H A Darch_timer.h41 asm volatile("msr cntp_ctl_el0, %0" : : "r" (val)); arch_timer_reg_write_cp15()
44 asm volatile("msr cntp_tval_el0, %0" : : "r" (val)); arch_timer_reg_write_cp15()
50 asm volatile("msr cntv_ctl_el0, %0" : : "r" (val)); arch_timer_reg_write_cp15()
53 asm volatile("msr cntv_tval_el0, %0" : : "r" (val)); arch_timer_reg_write_cp15()
105 asm volatile("msr cntkctl_el1, %0" : : "r" (cntkctl)); arch_timer_set_cntkctl()
H A Dfpsimdmacros.h51 msr fpcr, \state
74 msr fpsr, x\tmpnr
109 msr fpsr, x\tmpnr1
H A Dmmu_context.h42 " msr contextidr_el1, %0\n" contextidr_thread_switch()
61 " msr ttbr0_el1, %0 // set TTBR0\n" cpu_set_reserved_ttbr0()
88 " msr tcr_el1, %0 ;" __cpu_set_tcr_t0sz()
H A Dhw_breakpoint.h101 asm volatile("msr dbg" REG #N "_el1, %0" :: "r" (VAL));\
/linux-4.1.27/arch/x86/include/asm/
H A Dmsr.h4 #include <uapi/asm/msr.h>
12 struct msr { struct
24 struct msr reg;
25 struct msr *msrs;
60 static inline unsigned long long native_read_msr(unsigned int msr) native_read_msr() argument
64 asm volatile("rdmsr" : EAX_EDX_RET(val, low, high) : "c" (msr)); native_read_msr()
68 static inline unsigned long long native_read_msr_safe(unsigned int msr, native_read_msr_safe() argument
80 : "c" (msr), [fault] "i" (-EIO)); native_read_msr_safe()
84 static inline void native_write_msr(unsigned int msr, native_write_msr() argument
87 asm volatile("wrmsr" : : "c" (msr), "a"(low), "d" (high) : "memory"); native_write_msr()
91 notrace static inline int native_write_msr_safe(unsigned int msr, native_write_msr_safe() argument
102 : "c" (msr), "0" (low), "d" (high), native_write_msr_safe()
140 #define rdmsr(msr, low, high) \
142 u64 __val = native_read_msr((msr)); \
147 static inline void wrmsr(unsigned msr, unsigned low, unsigned high) wrmsr() argument
149 native_write_msr(msr, low, high); wrmsr()
152 #define rdmsrl(msr, val) \
153 ((val) = native_read_msr((msr)))
155 #define wrmsrl(msr, val) \
156 native_write_msr((msr), (u32)((u64)(val)), (u32)((u64)(val) >> 32))
159 static inline int wrmsr_safe(unsigned msr, unsigned low, unsigned high) wrmsr_safe() argument
161 return native_write_msr_safe(msr, low, high); wrmsr_safe()
165 #define rdmsr_safe(msr, low, high) \
168 u64 __val = native_read_msr_safe((msr), &__err); \
174 static inline int rdmsrl_safe(unsigned msr, unsigned long long *p) rdmsrl_safe() argument
178 *p = native_read_msr_safe(msr, &err); rdmsrl_safe()
208 #define wrmsrl_safe(msr, val) wrmsr_safe((msr), (u32)(val), \
215 struct msr *msrs_alloc(void);
216 void msrs_free(struct msr *msrs);
217 int msr_set_bit(u32 msr, u8 bit);
218 int msr_clear_bit(u32 msr, u8 bit);
225 void rdmsr_on_cpus(const struct cpumask *mask, u32 msr_no, struct msr *msrs);
226 void wrmsr_on_cpus(const struct cpumask *mask, u32 msr_no, struct msr *msrs);
255 struct msr *msrs) rdmsr_on_cpus()
260 struct msr *msrs) wrmsr_on_cpus()
H A Dmicrocode.h4 #define native_rdmsr(msr, val1, val2) \
6 u64 __val = native_read_msr((msr)); \
11 #define native_wrmsr(msr, low, high) \
12 native_write_msr(msr, low, high)
14 #define native_wrmsrl(msr, val) \
15 native_write_msr((msr), \
H A Dparavirt.h126 static inline u64 paravirt_read_msr(unsigned msr, int *err) paravirt_read_msr() argument
128 return PVOP_CALL2(u64, pv_cpu_ops.read_msr, msr, err); paravirt_read_msr()
131 static inline int paravirt_write_msr(unsigned msr, unsigned low, unsigned high) paravirt_write_msr() argument
133 return PVOP_CALL3(int, pv_cpu_ops.write_msr, msr, low, high); paravirt_write_msr()
137 #define rdmsr(msr, val1, val2) \
140 u64 _l = paravirt_read_msr(msr, &_err); \
145 #define wrmsr(msr, val1, val2) \
147 paravirt_write_msr(msr, val1, val2); \
150 #define rdmsrl(msr, val) \
153 val = paravirt_read_msr(msr, &_err); \
156 #define wrmsrl(msr, val) wrmsr(msr, (u32)((u64)(val)), ((u64)(val))>>32)
157 #define wrmsr_safe(msr, a, b) paravirt_write_msr(msr, a, b)
160 #define rdmsr_safe(msr, a, b) \
163 u64 _l = paravirt_read_msr(msr, &_err); \
169 static inline int rdmsrl_safe(unsigned msr, unsigned long long *p) rdmsrl_safe() argument
173 *p = paravirt_read_msr(msr, &err); rdmsrl_safe()
H A Dapic.h14 #include <asm/msr.h>
111 u64 msr; apic_is_x2apic_enabled() local
113 if (rdmsrl_safe(MSR_IA32_APICBASE, &msr)) apic_is_x2apic_enabled()
115 return msr & X2APIC_ENABLE; apic_is_x2apic_enabled()
145 u64 msr; native_apic_msr_read() local
150 rdmsrl(APIC_BASE_MSR + (reg >> 4), msr); native_apic_msr_read()
151 return (u32)msr; native_apic_msr_read()
H A Dkvm_host.h31 #include <asm/msr-index.h>
722 int (*set_msr)(struct kvm_vcpu *vcpu, struct msr_data *msr);
942 int kvm_set_msr(struct kvm_vcpu *vcpu, struct msr_data *msr);
970 int kvm_get_msr_common(struct kvm_vcpu *vcpu, u32 msr, u64 *pdata);
971 int kvm_set_msr_common(struct kvm_vcpu *vcpu, struct msr_data *msr);
1062 static inline unsigned long read_msr(unsigned long msr) read_msr() argument
1066 rdmsrl(msr, value); read_msr()
1154 void kvm_define_shared_msr(unsigned index, u32 msr);
1177 bool kvm_pmu_msr(struct kvm_vcpu *vcpu, u32 msr);
1178 int kvm_pmu_get_msr(struct kvm_vcpu *vcpu, u32 msr, u64 *data);
H A Dperf_event.h158 * to indicate other overflow conditions in the PERF_GLOBAL_STATUS msr.
265 unsigned msr; member in struct:perf_guest_switch_msr
/linux-4.1.27/arch/powerpc/include/asm/
H A Drunlatch.h21 unsigned long msr = mfmsr(); \
24 if (msr & MSR_EE) \
33 unsigned long msr = mfmsr(); \
36 if (msr & MSR_EE) \
H A Dkgdb.h42 * 32 gpr, 32 fpr, nip, msr, link, ctr
55 /* 32 GPRs (8 bytes), nip, msr, ccr, link, ctr, xer, acc (8 bytes), spefscr*/
H A Dperf_event.h38 asm volatile("mfmsr %0" : "=r" ((regs)->msr)); \
H A Dprobes.h50 regs->msr |= MSR_SINGLESTEP; enable_single_step()
57 regs->msr &= ~MSR_CE; enable_single_step()
H A Dhw_irq.h176 unsigned long msr = mfmsr(); arch_local_irq_enable()
177 SET_MSR_EE(msr | MSR_EE); arch_local_irq_enable()
195 return !(regs->msr & MSR_EE); arch_irq_disabled_regs()
H A Dptrace.h105 #define user_mode(regs) ((((regs)->msr) >> MSR_PR_LG) & 0x1)
107 #define user_mode(regs) (((regs)->msr & MSR_PR) != 0)
/linux-4.1.27/arch/arm64/kvm/
H A Dhyp.S76 msr sp_el1, x22
77 msr elr_el1, x23
78 msr spsr_el1, x24
84 msr sp_el0, x19
85 msr elr_el2, x20 // EL1 PC
86 msr spsr_el2, x21 // EL1 pstate
444 msr vmpidr_el2, x4
445 msr csselr_el1, x5
446 msr sctlr_el1, x6
447 msr actlr_el1, x7
448 msr cpacr_el1, x8
449 msr ttbr0_el1, x9
450 msr ttbr1_el1, x10
451 msr tcr_el1, x11
452 msr esr_el1, x12
453 msr afsr0_el1, x13
454 msr afsr1_el1, x14
455 msr far_el1, x15
456 msr mair_el1, x16
457 msr vbar_el1, x17
458 msr contextidr_el1, x18
459 msr tpidr_el0, x19
460 msr tpidrro_el0, x20
461 msr tpidr_el1, x21
462 msr amair_el1, x22
463 msr cntkctl_el1, x23
464 msr par_el1, x24
465 msr mdscr_el1, x25
506 msr dbgbcr15_el1, x20
507 msr dbgbcr14_el1, x19
508 msr dbgbcr13_el1, x18
509 msr dbgbcr12_el1, x17
510 msr dbgbcr11_el1, x16
511 msr dbgbcr10_el1, x15
512 msr dbgbcr9_el1, x14
513 msr dbgbcr8_el1, x13
514 msr dbgbcr7_el1, x12
515 msr dbgbcr6_el1, x11
516 msr dbgbcr5_el1, x10
517 msr dbgbcr4_el1, x9
518 msr dbgbcr3_el1, x8
519 msr dbgbcr2_el1, x7
520 msr dbgbcr1_el1, x6
521 msr dbgbcr0_el1, x5
550 msr dbgbvr15_el1, x20
551 msr dbgbvr14_el1, x19
552 msr dbgbvr13_el1, x18
553 msr dbgbvr12_el1, x17
554 msr dbgbvr11_el1, x16
555 msr dbgbvr10_el1, x15
556 msr dbgbvr9_el1, x14
557 msr dbgbvr8_el1, x13
558 msr dbgbvr7_el1, x12
559 msr dbgbvr6_el1, x11
560 msr dbgbvr5_el1, x10
561 msr dbgbvr4_el1, x9
562 msr dbgbvr3_el1, x8
563 msr dbgbvr2_el1, x7
564 msr dbgbvr1_el1, x6
565 msr dbgbvr0_el1, x5
594 msr dbgwcr15_el1, x20
595 msr dbgwcr14_el1, x19
596 msr dbgwcr13_el1, x18
597 msr dbgwcr12_el1, x17
598 msr dbgwcr11_el1, x16
599 msr dbgwcr10_el1, x15
600 msr dbgwcr9_el1, x14
601 msr dbgwcr8_el1, x13
602 msr dbgwcr7_el1, x12
603 msr dbgwcr6_el1, x11
604 msr dbgwcr5_el1, x10
605 msr dbgwcr4_el1, x9
606 msr dbgwcr3_el1, x8
607 msr dbgwcr2_el1, x7
608 msr dbgwcr1_el1, x6
609 msr dbgwcr0_el1, x5
638 msr dbgwvr15_el1, x20
639 msr dbgwvr14_el1, x19
640 msr dbgwvr13_el1, x18
641 msr dbgwvr12_el1, x17
642 msr dbgwvr11_el1, x16
643 msr dbgwvr10_el1, x15
644 msr dbgwvr9_el1, x14
645 msr dbgwvr8_el1, x13
646 msr dbgwvr7_el1, x12
647 msr dbgwvr6_el1, x11
648 msr dbgwvr5_el1, x10
649 msr dbgwvr4_el1, x9
650 msr dbgwvr3_el1, x8
651 msr dbgwvr2_el1, x7
652 msr dbgwvr1_el1, x6
653 msr dbgwvr0_el1, x5
656 msr mdccint_el1, x21
737 msr spsr_abt, x4
738 msr spsr_und, x5
739 msr spsr_irq, x6
740 msr spsr_fiq, x7
745 msr dacr32_el2, x4
746 msr ifsr32_el2, x5
747 msr fpexc32_el2, x6
751 msr dbgvcr32_el2, x7
757 msr teecr32_el1, x4
758 msr teehbr32_el1, x5
764 msr hcr_el2, x2
766 msr cptr_el2, x2
769 msr hstr_el2, x2
782 msr mdcr_el2, x2
787 msr hcr_el2, x2
788 msr cptr_el2, xzr
789 msr hstr_el2, xzr
793 msr mdcr_el2, x2
800 msr vttbr_el2, x2
804 msr vttbr_el2, xzr
819 msr hcr_el2, x24
830 msr hcr_el2, x24
855 msr cntv_ctl_el0, xzr
860 msr cnthctl_el2, x2
863 msr cntvoff_el2, xzr
873 msr cnthctl_el2, x2
881 msr cntvoff_el2, x3
883 msr cntv_cval_el0, x2
888 msr cntv_ctl_el0, x2
929 msr tpidr_el2, x0 // Save the vcpu register
1013 msr vttbr_el2, x2
1034 msr vttbr_el2, xzr
1049 msr vttbr_el2, x2
1056 msr vttbr_el2, xzr
1106 msr spsr_el2, lr
1108 msr elr_el2, lr
1241 msr par_el1, x0
H A Dhyp-init.S62 msr ttbr0_el2, x0
88 msr tcr_el2, x4
97 msr vtcr_el2, x4
100 msr mair_el2, x4
111 msr sctlr_el2, x4
125 msr ttbr0_el2, x1
137 msr vbar_el2, x3
/linux-4.1.27/arch/microblaze/kernel/
H A Dprocess.c43 pr_info(" msr=%08lX, ear=%08lX, esr=%08lX, fsr=%08lX\n", show_regs()
44 regs->msr, regs->ear, regs->esr, regs->fsr); show_regs()
69 local_save_flags(childregs->msr); copy_thread()
71 ti->cpu_context.msr = childregs->msr & ~MSR_IE; copy_thread()
83 ti->cpu_context.msr = (unsigned long)childregs->msr; copy_thread()
85 childregs->msr |= MSR_UMS; copy_thread()
97 childregs->msr &= ~MSR_EIP; copy_thread()
98 childregs->msr |= MSR_IE; copy_thread()
99 childregs->msr &= ~MSR_VM; copy_thread()
100 childregs->msr |= MSR_VMS; copy_thread()
101 childregs->msr |= MSR_EE; /* exceptions will be enabled*/ copy_thread()
103 ti->cpu_context.msr = (childregs->msr|MSR_VM); copy_thread()
104 ti->cpu_context.msr &= ~MSR_UMS; /* switch_to to kernel mode */ copy_thread()
105 ti->cpu_context.msr &= ~MSR_IE; copy_thread()
149 regs->msr |= MSR_UMS; start_thread()
150 regs->msr &= ~MSR_VM; start_thread()
H A Dsetup.c100 unsigned int fdt, unsigned int msr, unsigned int tlb0, machine_early_init()
167 if (msr) { machine_early_init()
169 pr_cont("CPU don't have it %x\n", msr); machine_early_init()
172 if (!msr) { machine_early_init()
174 pr_cont("CPU have it %x\n", msr); machine_early_init()
99 machine_early_init(const char *cmdline, unsigned int ram, unsigned int fdt, unsigned int msr, unsigned int tlb0, unsigned int tlb1) machine_early_init() argument
H A Dkgdb.c43 /* registers r0 - r31, pc, msr, ear, esr, fsr + do not save pt_mode */ pt_regs_to_gdb_regs()
99 /* registers r0 - r31, pc, msr, ear, esr, fsr + do not save pt_mode */ sleeping_thread_to_gdb_regs()
H A Dasm-offsets.c25 DEFINE(PT_MSR, offsetof(struct pt_regs, msr)); main()
121 DEFINE(CC_MSR, offsetof(struct cpu_context, msr)); main()
H A Dhead.S75 * r8 == 0 - msr instructions are implemented
76 * r8 != 0 - msr instructions are not implemented
79 msrclr r8, 0 /* clear nothing - just read msr for test */
80 cmpu r8, r8, r1 /* r1 must contain msr reg content */
/linux-4.1.27/arch/x86/include/uapi/asm/
H A Dmsr.h4 #include <asm/msr-index.h>
H A Dsvm.h116 { SVM_EXIT_MSR, "msr" }, \
H A Dprocessor-flags.h130 * AMD and Transmeta use MSRs for configuration; see <asm/msr-index.h>
/linux-4.1.27/arch/x86/kernel/
H A Dtrace_clock.c6 #include <asm/msr.h>
H A Dmsr.c21 * This driver uses /dev/cpu/%d/msr where %d is the minor number, and on
44 #include <asm/msr.h>
210 "msr%d", cpu); msr_device_create()
244 return kasprintf(GFP_KERNEL, "cpu/%u/msr", MINOR(dev->devt)); msr_devnode()
252 if (__register_chrdev(MSR_MAJOR, 0, NR_CPUS, "cpu/msr", &msr_fops)) { msr_init()
253 pr_err("unable to get major %d for msr\n", MSR_MAJOR); msr_init()
257 msr_class = class_create(THIS_MODULE, "msr"); msr_init()
283 __unregister_chrdev(MSR_MAJOR, 0, NR_CPUS, "cpu/msr");
296 __unregister_chrdev(MSR_MAJOR, 0, NR_CPUS, "cpu/msr"); msr_exit()
H A Damd_nb.c143 u64 base, msr; amd_get_mmconfig_range() local
154 rdmsrl(address, msr); amd_get_mmconfig_range()
157 if (!(msr & FAM10H_MMIO_CONF_ENABLE)) amd_get_mmconfig_range()
160 base = msr & (FAM10H_MMIO_CONF_BASE_MASK<<FAM10H_MMIO_CONF_BASE_SHIFT); amd_get_mmconfig_range()
162 segn_busn_bits = (msr >> FAM10H_MMIO_CONF_BUSRANGE_SHIFT) & amd_get_mmconfig_range()
H A Dreboot_fixups_32.c14 #include <asm/msr.h>
H A Dkvmclock.c22 #include <asm/msr.h>
168 printk(KERN_INFO "kvm-clock: cpu %d, msr %x:%x, %s\n", kvm_register_clock()
200 * that does not have the 'enable' bit set in the msr
H A DMakefile58 obj-$(CONFIG_X86_MSR) += msr.o
H A Dcpuid.c45 #include <asm/msr.h>
H A Dmmconf-fam10h_64.c15 #include <asm/msr.h>
H A Dverify_cpu.S34 #include <asm/msr-index.h>
/linux-4.1.27/tools/power/cpupower/debug/i386/
H A Dcentrino-decode.c29 static int rdmsr(unsigned int cpu, unsigned int msr, rdmsr() argument
42 sprintf(file, "/dev/cpu/%d/msr", cpu); rdmsr()
48 if (lseek(fd, msr, SEEK_CUR) == -1) rdmsr()
64 static void decode (unsigned int msr) decode() argument
69 multiplier = ((msr >> 8) & 0xFF); decode()
71 mv = (((msr & 0xFF) * 16) + 700); decode()
73 printf("0x%x means multiplier %d @ %d mV\n", msr, multiplier, mv); decode()
86 printf("or you are not root, or the msr driver is not present\n"); decode_live()
H A Dpowernow-k8-decode.c31 uint64_t msr = 0; get_fidvid() local
38 sprintf(file, "/dev/cpu/%d/msr", cpu); get_fidvid()
44 if (read(fd, &msr, 8) != 8) get_fidvid()
47 *fid = ((uint32_t )(msr & 0xffffffffull)) & MSR_S_LO_CURRENT_FID; get_fidvid()
48 *vid = ((uint32_t )(msr>>32 & 0xffffffffull)) & MSR_S_HI_CURRENT_VID; get_fidvid()
85 printf("or you are not root, or the msr driver is not present\n"); main()
/linux-4.1.27/tools/power/x86/turbostat/
H A Dturbostat.c261 int get_msr(int cpu, off_t offset, unsigned long long *msr) get_msr() argument
267 sprintf(pathname, "/dev/cpu/%d/msr", cpu); get_msr()
270 err(-1, "%s open failed, try chown or chmod +r /dev/cpu/*/msr, or run as root", pathname); get_msr()
272 retval = pread(fd, msr, sizeof *msr, offset); get_msr()
275 if (retval != sizeof *msr) get_msr()
525 /* msr */ format_counters()
973 unsigned long long msr; get_counters() local
990 if (get_msr(cpu, MSR_SMI_COUNT, &msr)) get_counters()
992 t->smi_count = msr & 0xFFFFFFFF; get_counters()
995 if (get_msr(cpu, extra_delta_offset32, &msr)) get_counters()
997 t->extra_delta32 = msr & 0xFFFFFFFF; get_counters()
1005 if (get_msr(cpu, extra_msr_offset32, &msr)) get_counters()
1007 t->extra_msr32 = msr & 0xFFFFFFFF; get_counters()
1041 if (get_msr(cpu, MSR_IA32_THERM_STATUS, &msr)) get_counters()
1043 c->core_temp_c = tcc_activation_temp - ((msr >> 16) & 0x7F); get_counters()
1082 if (get_msr(cpu, MSR_PKG_ENERGY_STATUS, &msr)) get_counters()
1084 p->energy_pkg = msr & 0xFFFFFFFF; get_counters()
1087 if (get_msr(cpu, MSR_PP0_ENERGY_STATUS, &msr)) get_counters()
1089 p->energy_cores = msr & 0xFFFFFFFF; get_counters()
1092 if (get_msr(cpu, MSR_DRAM_ENERGY_STATUS, &msr)) get_counters()
1094 p->energy_dram = msr & 0xFFFFFFFF; get_counters()
1097 if (get_msr(cpu, MSR_PP1_ENERGY_STATUS, &msr)) get_counters()
1099 p->energy_gfx = msr & 0xFFFFFFFF; get_counters()
1102 if (get_msr(cpu, MSR_PKG_PERF_STATUS, &msr)) get_counters()
1104 p->rapl_pkg_perf_status = msr & 0xFFFFFFFF; get_counters()
1107 if (get_msr(cpu, MSR_DRAM_PERF_STATUS, &msr)) get_counters()
1109 p->rapl_dram_perf_status = msr & 0xFFFFFFFF; get_counters()
1112 if (get_msr(cpu, MSR_IA32_PACKAGE_THERM_STATUS, &msr)) get_counters()
1114 p->pkg_temp_c = tcc_activation_temp - ((msr >> 16) & 0x7F); get_counters()
1155 unsigned long long msr; dump_nhm_platform_info() local
1158 get_msr(base_cpu, MSR_NHM_PLATFORM_INFO, &msr); dump_nhm_platform_info()
1160 fprintf(stderr, "cpu0: MSR_NHM_PLATFORM_INFO: 0x%08llx\n", msr); dump_nhm_platform_info()
1162 ratio = (msr >> 40) & 0xFF; dump_nhm_platform_info()
1166 ratio = (msr >> 8) & 0xFF; dump_nhm_platform_info()
1170 get_msr(base_cpu, MSR_IA32_POWER_CTL, &msr); dump_nhm_platform_info()
1172 msr, msr & 0x2 ? "EN" : "DIS"); dump_nhm_platform_info()
1180 unsigned long long msr; dump_hsw_turbo_ratio_limits() local
1183 get_msr(base_cpu, MSR_TURBO_RATIO_LIMIT2, &msr); dump_hsw_turbo_ratio_limits()
1185 fprintf(stderr, "cpu0: MSR_TURBO_RATIO_LIMIT2: 0x%08llx\n", msr); dump_hsw_turbo_ratio_limits()
1187 ratio = (msr >> 8) & 0xFF; dump_hsw_turbo_ratio_limits()
1192 ratio = (msr >> 0) & 0xFF; dump_hsw_turbo_ratio_limits()
1202 unsigned long long msr; dump_ivt_turbo_ratio_limits() local
1205 get_msr(base_cpu, MSR_TURBO_RATIO_LIMIT1, &msr); dump_ivt_turbo_ratio_limits()
1207 fprintf(stderr, "cpu0: MSR_TURBO_RATIO_LIMIT1: 0x%08llx\n", msr); dump_ivt_turbo_ratio_limits()
1209 ratio = (msr >> 56) & 0xFF; dump_ivt_turbo_ratio_limits()
1214 ratio = (msr >> 48) & 0xFF; dump_ivt_turbo_ratio_limits()
1219 ratio = (msr >> 40) & 0xFF; dump_ivt_turbo_ratio_limits()
1224 ratio = (msr >> 32) & 0xFF; dump_ivt_turbo_ratio_limits()
1229 ratio = (msr >> 24) & 0xFF; dump_ivt_turbo_ratio_limits()
1234 ratio = (msr >> 16) & 0xFF; dump_ivt_turbo_ratio_limits()
1239 ratio = (msr >> 8) & 0xFF; dump_ivt_turbo_ratio_limits()
1244 ratio = (msr >> 0) & 0xFF; dump_ivt_turbo_ratio_limits()
1254 unsigned long long msr; dump_nhm_turbo_ratio_limits() local
1257 get_msr(base_cpu, MSR_TURBO_RATIO_LIMIT, &msr); dump_nhm_turbo_ratio_limits()
1259 fprintf(stderr, "cpu0: MSR_TURBO_RATIO_LIMIT: 0x%08llx\n", msr); dump_nhm_turbo_ratio_limits()
1261 ratio = (msr >> 56) & 0xFF; dump_nhm_turbo_ratio_limits()
1266 ratio = (msr >> 48) & 0xFF; dump_nhm_turbo_ratio_limits()
1271 ratio = (msr >> 40) & 0xFF; dump_nhm_turbo_ratio_limits()
1276 ratio = (msr >> 32) & 0xFF; dump_nhm_turbo_ratio_limits()
1281 ratio = (msr >> 24) & 0xFF; dump_nhm_turbo_ratio_limits()
1286 ratio = (msr >> 16) & 0xFF; dump_nhm_turbo_ratio_limits()
1291 ratio = (msr >> 8) & 0xFF; dump_nhm_turbo_ratio_limits()
1296 ratio = (msr >> 0) & 0xFF; dump_nhm_turbo_ratio_limits()
1308 unsigned long long msr; dump_knl_turbo_ratio_limits() local
1313 get_msr(base_cpu, MSR_NHM_TURBO_RATIO_LIMIT, &msr); dump_knl_turbo_ratio_limits()
1316 msr); dump_knl_turbo_ratio_limits()
1339 cores = msr & 0xFF; dump_knl_turbo_ratio_limits()
1340 ratio = (msr >> 8) && 0xFF; dump_knl_turbo_ratio_limits()
1347 delta_cores = (msr >> i) & 0x1F; dump_knl_turbo_ratio_limits()
1348 delta_ratio = (msr >> (i + 5)) && 0x7; dump_knl_turbo_ratio_limits()
1367 unsigned long long msr; dump_nhm_cst_cfg() local
1369 get_msr(base_cpu, MSR_NHM_SNB_PKG_CST_CFG_CTL, &msr); dump_nhm_cst_cfg()
1374 fprintf(stderr, "cpu0: MSR_NHM_SNB_PKG_CST_CFG_CTL: 0x%08llx", msr); dump_nhm_cst_cfg()
1377 (msr & SNB_C3_AUTO_UNDEMOTE) ? "UNdemote-C3, " : "", dump_nhm_cst_cfg()
1378 (msr & SNB_C1_AUTO_UNDEMOTE) ? "UNdemote-C1, " : "", dump_nhm_cst_cfg()
1379 (msr & NHM_C3_AUTO_DEMOTE) ? "demote-C3, " : "", dump_nhm_cst_cfg()
1380 (msr & NHM_C1_AUTO_DEMOTE) ? "demote-C1, " : "", dump_nhm_cst_cfg()
1381 (msr & (1 << 15)) ? "" : "UN", dump_nhm_cst_cfg()
1382 (unsigned int)msr & 7, dump_nhm_cst_cfg()
1700 sprintf(pathname, "/dev/cpu/%d/msr", base_cpu); check_dev_msr()
1702 if (system("/sbin/modprobe msr > /dev/null 2>&1")) check_dev_msr()
1703 err(-5, "no /dev/cpu/0/msr, Try \"# modprobe msr\" "); check_dev_msr()
1729 sprintf(pathname, "/dev/cpu/%d/msr", base_cpu); check_permissions()
1732 warn("/dev/cpu/0/msr open failed, try chown or chmod +r /dev/cpu/*/msr"); check_permissions()
1762 unsigned long long msr; probe_nhm_msrs() local
1812 get_msr(base_cpu, MSR_NHM_SNB_PKG_CST_CFG_CTL, &msr); probe_nhm_msrs()
1814 pkg_cstate_limit = pkg_cstate_limits[msr & 0xF]; probe_nhm_msrs()
1906 unsigned long long msr; print_epb() local
1924 if (get_msr(cpu, MSR_IA32_ENERGY_PERF_BIAS, &msr)) print_epb()
1927 switch (msr & 0xF) { print_epb()
1941 fprintf(stderr, "cpu%d: MSR_IA32_ENERGY_PERF_BIAS: 0x%08llx (%s)\n", cpu, msr, epb_string); print_epb()
1951 unsigned long long msr; print_perf_limit() local
1966 get_msr(cpu, MSR_CORE_PERF_LIMIT_REASONS, &msr); print_perf_limit()
1967 fprintf(stderr, "cpu%d: MSR_CORE_PERF_LIMIT_REASONS, 0x%08llx", cpu, msr); print_perf_limit()
1969 (msr & 1 << 15) ? "bit15, " : "", print_perf_limit()
1970 (msr & 1 << 14) ? "bit14, " : "", print_perf_limit()
1971 (msr & 1 << 13) ? "Transitions, " : "", print_perf_limit()
1972 (msr & 1 << 12) ? "MultiCoreTurbo, " : "", print_perf_limit()
1973 (msr & 1 << 11) ? "PkgPwrL2, " : "", print_perf_limit()
1974 (msr & 1 << 10) ? "PkgPwrL1, " : "", print_perf_limit()
1975 (msr & 1 << 9) ? "CorePwr, " : "", print_perf_limit()
1976 (msr & 1 << 8) ? "Amps, " : "", print_perf_limit()
1977 (msr & 1 << 6) ? "VR-Therm, " : "", print_perf_limit()
1978 (msr & 1 << 5) ? "Auto-HWP, " : "", print_perf_limit()
1979 (msr & 1 << 4) ? "Graphics, " : "", print_perf_limit()
1980 (msr & 1 << 2) ? "bit2, " : "", print_perf_limit()
1981 (msr & 1 << 1) ? "ThermStatus, " : "", print_perf_limit()
1982 (msr & 1 << 0) ? "PROCHOT, " : ""); print_perf_limit()
1984 (msr & 1 << 31) ? "bit31, " : "", print_perf_limit()
1985 (msr & 1 << 30) ? "bit30, " : "", print_perf_limit()
1986 (msr & 1 << 29) ? "Transitions, " : "", print_perf_limit()
1987 (msr & 1 << 28) ? "MultiCoreTurbo, " : "", print_perf_limit()
1988 (msr & 1 << 27) ? "PkgPwrL2, " : "", print_perf_limit()
1989 (msr & 1 << 26) ? "PkgPwrL1, " : "", print_perf_limit()
1990 (msr & 1 << 25) ? "CorePwr, " : "", print_perf_limit()
1991 (msr & 1 << 24) ? "Amps, " : "", print_perf_limit()
1992 (msr & 1 << 22) ? "VR-Therm, " : "", print_perf_limit()
1993 (msr & 1 << 21) ? "Auto-HWP, " : "", print_perf_limit()
1994 (msr & 1 << 20) ? "Graphics, " : "", print_perf_limit()
1995 (msr & 1 << 18) ? "bit18, " : "", print_perf_limit()
1996 (msr & 1 << 17) ? "ThermStatus, " : "", print_perf_limit()
1997 (msr & 1 << 16) ? "PROCHOT, " : ""); print_perf_limit()
2001 get_msr(cpu, MSR_GFX_PERF_LIMIT_REASONS, &msr); print_perf_limit()
2002 fprintf(stderr, "cpu%d: MSR_GFX_PERF_LIMIT_REASONS, 0x%08llx", cpu, msr); print_perf_limit()
2004 (msr & 1 << 0) ? "PROCHOT, " : "", print_perf_limit()
2005 (msr & 1 << 1) ? "ThermStatus, " : "", print_perf_limit()
2006 (msr & 1 << 4) ? "Graphics, " : "", print_perf_limit()
2007 (msr & 1 << 6) ? "VR-Therm, " : "", print_perf_limit()
2008 (msr & 1 << 8) ? "Amps, " : "", print_perf_limit()
2009 (msr & 1 << 9) ? "GFXPwr, " : "", print_perf_limit()
2010 (msr & 1 << 10) ? "PkgPwrL1, " : "", print_perf_limit()
2011 (msr & 1 << 11) ? "PkgPwrL2, " : ""); print_perf_limit()
2013 (msr & 1 << 16) ? "PROCHOT, " : "", print_perf_limit()
2014 (msr & 1 << 17) ? "ThermStatus, " : "", print_perf_limit()
2015 (msr & 1 << 20) ? "Graphics, " : "", print_perf_limit()
2016 (msr & 1 << 22) ? "VR-Therm, " : "", print_perf_limit()
2017 (msr & 1 << 24) ? "Amps, " : "", print_perf_limit()
2018 (msr & 1 << 25) ? "GFXPwr, " : "", print_perf_limit()
2019 (msr & 1 << 26) ? "PkgPwrL1, " : "", print_perf_limit()
2020 (msr & 1 << 27) ? "PkgPwrL2, " : ""); print_perf_limit()
2023 get_msr(cpu, MSR_RING_PERF_LIMIT_REASONS, &msr); print_perf_limit()
2024 fprintf(stderr, "cpu%d: MSR_RING_PERF_LIMIT_REASONS, 0x%08llx", cpu, msr); print_perf_limit()
2026 (msr & 1 << 0) ? "PROCHOT, " : "", print_perf_limit()
2027 (msr & 1 << 1) ? "ThermStatus, " : "", print_perf_limit()
2028 (msr & 1 << 6) ? "VR-Therm, " : "", print_perf_limit()
2029 (msr & 1 << 8) ? "Amps, " : "", print_perf_limit()
2030 (msr & 1 << 10) ? "PkgPwrL1, " : "", print_perf_limit()
2031 (msr & 1 << 11) ? "PkgPwrL2, " : ""); print_perf_limit()
2033 (msr & 1 << 16) ? "PROCHOT, " : "", print_perf_limit()
2034 (msr & 1 << 17) ? "ThermStatus, " : "", print_perf_limit()
2035 (msr & 1 << 22) ? "VR-Therm, " : "", print_perf_limit()
2036 (msr & 1 << 24) ? "Amps, " : "", print_perf_limit()
2037 (msr & 1 << 26) ? "PkgPwrL1, " : "", print_perf_limit()
2038 (msr & 1 << 27) ? "PkgPwrL2, " : ""); print_perf_limit()
2048 unsigned long long msr; get_tdp() local
2051 if (!get_msr(base_cpu, MSR_PKG_POWER_INFO, &msr)) get_tdp()
2052 return ((msr >> 0) & RAPL_POWER_GRANULARITY) * rapl_power_units; get_tdp()
2091 unsigned long long msr; rapl_probe() local
2134 if (get_msr(base_cpu, MSR_RAPL_POWER_UNIT, &msr)) rapl_probe()
2137 rapl_power_units = 1.0 / (1 << (msr & 0xF)); rapl_probe()
2139 rapl_energy_units = 1.0 * (1 << (msr >> 8 & 0x1F)) / 1000000; rapl_probe()
2141 rapl_energy_units = 1.0 / (1 << (msr >> 8 & 0x1F)); rapl_probe()
2145 time_unit = msr >> 16 & 0xF; rapl_probe()
2183 unsigned long long msr; print_thermal() local
2202 if (get_msr(cpu, MSR_IA32_PACKAGE_THERM_STATUS, &msr)) print_thermal()
2205 dts = (msr >> 16) & 0x7F; print_thermal()
2207 cpu, msr, tcc_activation_temp - dts); print_thermal()
2210 if (get_msr(cpu, MSR_IA32_PACKAGE_THERM_INTERRUPT, &msr)) print_thermal()
2213 dts = (msr >> 16) & 0x7F; print_thermal()
2214 dts2 = (msr >> 8) & 0x7F; print_thermal()
2216 cpu, msr, tcc_activation_temp - dts, tcc_activation_temp - dts2); print_thermal()
2224 if (get_msr(cpu, MSR_IA32_THERM_STATUS, &msr)) print_thermal()
2227 dts = (msr >> 16) & 0x7F; print_thermal()
2228 resolution = (msr >> 27) & 0xF; print_thermal()
2230 cpu, msr, tcc_activation_temp - dts, resolution); print_thermal()
2233 if (get_msr(cpu, MSR_IA32_THERM_INTERRUPT, &msr)) print_thermal()
2236 dts = (msr >> 16) & 0x7F; print_thermal()
2237 dts2 = (msr >> 8) & 0x7F; print_thermal()
2239 cpu, msr, tcc_activation_temp - dts, tcc_activation_temp - dts2); print_thermal()
2246 void print_power_limit_msr(int cpu, unsigned long long msr, char *label) print_power_limit_msr() argument
2250 ((msr >> 15) & 1) ? "EN" : "DIS", print_power_limit_msr()
2251 ((msr >> 0) & 0x7FFF) * rapl_power_units, print_power_limit_msr()
2252 (1.0 + (((msr >> 22) & 0x3)/4.0)) * (1 << ((msr >> 17) & 0x1F)) * rapl_time_units, print_power_limit_msr()
2253 (((msr >> 16) & 1) ? "EN" : "DIS")); print_power_limit_msr()
2260 unsigned long long msr; print_rapl() local
2276 if (get_msr(cpu, MSR_RAPL_POWER_UNIT, &msr)) print_rapl()
2281 "(%f Watts, %f Joules, %f sec.)\n", cpu, msr, print_rapl()
2286 if (get_msr(cpu, MSR_PKG_POWER_INFO, &msr)) print_rapl()
2291 cpu, msr, print_rapl()
2292 ((msr >> 0) & RAPL_POWER_GRANULARITY) * rapl_power_units, print_rapl()
2293 ((msr >> 16) & RAPL_POWER_GRANULARITY) * rapl_power_units, print_rapl()
2294 ((msr >> 32) & RAPL_POWER_GRANULARITY) * rapl_power_units, print_rapl()
2295 ((msr >> 48) & RAPL_TIME_GRANULARITY) * rapl_time_units); print_rapl()
2300 if (get_msr(cpu, MSR_PKG_POWER_LIMIT, &msr)) print_rapl()
2304 cpu, msr, (msr >> 63) & 1 ? "": "UN"); print_rapl()
2306 print_power_limit_msr(cpu, msr, "PKG Limit #1"); print_rapl()
2309 ((msr >> 47) & 1) ? "EN" : "DIS", print_rapl()
2310 ((msr >> 32) & 0x7FFF) * rapl_power_units, print_rapl()
2311 (1.0 + (((msr >> 54) & 0x3)/4.0)) * (1 << ((msr >> 49) & 0x1F)) * rapl_time_units, print_rapl()
2312 ((msr >> 48) & 1) ? "EN" : "DIS"); print_rapl()
2316 if (get_msr(cpu, MSR_DRAM_POWER_INFO, &msr)) print_rapl()
2320 cpu, msr, print_rapl()
2321 ((msr >> 0) & RAPL_POWER_GRANULARITY) * rapl_power_units, print_rapl()
2322 ((msr >> 16) & RAPL_POWER_GRANULARITY) * rapl_power_units, print_rapl()
2323 ((msr >> 32) & RAPL_POWER_GRANULARITY) * rapl_power_units, print_rapl()
2324 ((msr >> 48) & RAPL_TIME_GRANULARITY) * rapl_time_units); print_rapl()
2327 if (get_msr(cpu, MSR_DRAM_POWER_LIMIT, &msr)) print_rapl()
2330 cpu, msr, (msr >> 31) & 1 ? "": "UN"); print_rapl()
2332 print_power_limit_msr(cpu, msr, "DRAM Limit"); print_rapl()
2336 if (get_msr(cpu, MSR_PP0_POLICY, &msr)) print_rapl()
2339 fprintf(stderr, "cpu%d: MSR_PP0_POLICY: %lld\n", cpu, msr & 0xF); print_rapl()
2345 if (get_msr(cpu, MSR_PP0_POWER_LIMIT, &msr)) print_rapl()
2348 cpu, msr, (msr >> 31) & 1 ? "": "UN"); print_rapl()
2349 print_power_limit_msr(cpu, msr, "Cores Limit"); print_rapl()
2354 if (get_msr(cpu, MSR_PP1_POLICY, &msr)) print_rapl()
2357 fprintf(stderr, "cpu%d: MSR_PP1_POLICY: %lld\n", cpu, msr & 0xF); print_rapl()
2359 if (get_msr(cpu, MSR_PP1_POWER_LIMIT, &msr)) print_rapl()
2362 cpu, msr, (msr >> 31) & 1 ? "": "UN"); print_rapl()
2363 print_power_limit_msr(cpu, msr, "GFX Limit"); print_rapl()
2475 unsigned long long msr = 3; slm_bclk() local
2479 if (get_msr(base_cpu, MSR_FSB_FREQ, &msr)) slm_bclk()
2482 i = msr & 0xf; slm_bclk()
2485 msr = 3; slm_bclk()
2518 unsigned long long msr; set_temperature_target() local
2547 if (get_msr(base_cpu, MSR_IA32_TEMPERATURE_TARGET, &msr)) set_temperature_target()
2550 target_c_local = (msr >> 16) & 0xFF; set_temperature_target()
2554 cpu, msr, target_c_local); set_temperature_target()
2704 "--counter msr print 32-bit counter at address \"msr\"\n" help()
2705 "--Counter msr print 64-bit Counter at address \"msr\"\n" help()
2706 "--msr msr print 32-bit value at address \"msr\"\n" help()
2707 "--MSR msr print 64-bit Value at address \"msr\"\n" help()
3034 {"msr", required_argument, 0, 'm'}, cmdline()
H A DMakefile12 CFLAGS += -DMSRHEADER='"../../../../arch/x86/include/uapi/asm/msr-index.h"'
/linux-4.1.27/arch/m68k/bvme6000/
H A Dconfig.c164 unsigned char msr = rtc->msr & 0xc0; bvme6000_timer_int() local
166 rtc->msr = msr | 0x20; /* Ack the interrupt */ bvme6000_timer_int()
183 unsigned char msr = rtc->msr & 0xc0; bvme6000_sched_init() local
185 rtc->msr = 0; /* Ensure timer registers accessible */ bvme6000_sched_init()
196 rtc->msr = 0x40; /* Access int.cntrl, etc */ bvme6000_sched_init()
201 rtc->msr = 0; /* Access timer 1 control */ bvme6000_sched_init()
204 rtc->msr = msr; bvme6000_sched_init()
225 unsigned char msr = rtc->msr & 0xc0; bvme6000_gettimeoffset() local
229 rtc->msr = 0; /* Ensure timer registers accessible */ bvme6000_gettimeoffset()
233 t1int = rtc->msr & 0x20; bvme6000_gettimeoffset()
238 } while (t1int != (rtc->msr & 0x20) || bvme6000_gettimeoffset()
249 rtc->msr = msr; bvme6000_gettimeoffset()
272 unsigned char msr = rtc->msr & 0xc0; bvme6000_hwclk() local
274 rtc->msr = 0x40; /* Ensure clock and real-time-mode-register bvme6000_hwclk()
305 rtc->msr = msr; bvme6000_hwclk()
323 unsigned char msr = rtc->msr & 0xc0; bvme6000_set_clock_mmss() local
327 rtc->msr = 0; /* Ensure clock accessible */ bvme6000_set_clock_mmss()
347 rtc->msr = msr; bvme6000_set_clock_mmss()
H A Drtc.c41 unsigned char msr; rtc_ioctl() local
51 msr = rtc->msr & 0xc0; rtc_ioctl()
52 rtc->msr = 0x40; rtc_ioctl()
65 rtc->msr = msr; rtc_ioctl()
107 msr = rtc->msr & 0xc0; rtc_ioctl()
108 rtc->msr = 0x40; rtc_ioctl()
122 rtc->msr = msr; rtc_ioctl()
/linux-4.1.27/arch/arm/include/asm/
H A Dirqflags.h66 " msr cpsr_c, %1" arch_local_irq_save()
82 " msr cpsr_c, %0" arch_local_irq_enable()
97 " msr cpsr_c, %0" arch_local_irq_disable()
112 " msr cpsr_c, %0" \
127 " msr cpsr_c, %0" \
153 " msr " IRQMASK_REG_NAME_W ", %0 @ local_irq_restore" arch_local_irq_restore()
H A Dassembler.h103 msr cpsr_c, #PSR_I_BIT | SVC_MODE
107 msr cpsr_c, #SVC_MODE
168 msr primask, \oldcpsr
170 msr cpsr_c, \oldcpsr
304 msr cpsr_c, \reg
308 msr cpsr_c, #\mode
330 msr spsr_cxsf, \reg
333 1: msr cpsr_c, \reg
/linux-4.1.27/arch/arm64/mm/
H A Dproc.S56 msr sctlr_el1, x0
74 msr sctlr_el1, x1 // disable the MMU
152 msr tpidr_el0, x2
153 msr tpidrro_el0, x3
154 msr contextidr_el1, x4
155 msr mair_el1, x5
156 msr cpacr_el1, x6
157 msr ttbr0_el1, x1
158 msr ttbr1_el1, x7
160 msr tcr_el1, x8
161 msr vbar_el1, x9
162 msr mdscr_el1, x10
167 msr oslar_el1, x11
186 msr ttbr0_el1, x0 // set TTBR0
205 msr cpacr_el1, x0 // Enable FP/ASIMD
207 msr mdscr_el1, x0 // access to the DCC from EL0
225 msr mair_el1, x5
248 msr tcr_el1, x10
H A Dcache.S50 msr csselr_el1, x10 // select current cache level in csselr
79 msr csselr_el1, x10 // select current cache level in csselr
/linux-4.1.27/arch/powerpc/kernel/
H A Dsignal_64.c95 unsigned long msr = regs->msr; setup_sigcontext() local
110 msr |= MSR_VEC; setup_sigcontext()
129 msr &= ~MSR_VSX; setup_sigcontext()
143 msr |= MSR_VSX; setup_sigcontext()
149 err |= __put_user(msr, &sc->gp_regs[PT_MSR]); setup_sigcontext()
189 unsigned long msr = regs->msr; setup_tm_sigcontexts() local
192 BUG_ON(!MSR_TM_ACTIVE(regs->msr)); setup_tm_sigcontexts()
199 regs->msr &= ~MSR_TS_MASK; setup_tm_sigcontexts()
216 if (msr & MSR_VEC) setup_tm_sigcontexts()
228 msr |= MSR_VEC; setup_tm_sigcontexts()
236 if (msr & MSR_VEC) setup_tm_sigcontexts()
250 if (msr & MSR_FP) setup_tm_sigcontexts()
268 if (msr & MSR_VSX) setup_tm_sigcontexts()
276 msr |= MSR_VSX; setup_tm_sigcontexts()
286 err |= __put_user(msr, &tm_sc->gp_regs[PT_MSR]); setup_tm_sigcontexts()
287 err |= __put_user(msr, &sc->gp_regs[PT_MSR]); setup_tm_sigcontexts()
309 unsigned long msr; restore_sigcontext() local
322 err |= __get_user(msr, &sc->gp_regs[PT_MSR]); restore_sigcontext()
324 regs->msr = (regs->msr & ~MSR_LE) | (msr & MSR_LE); restore_sigcontext()
355 regs->msr &= ~(MSR_FP | MSR_FE0 | MSR_FE1 | MSR_VEC | MSR_VSX); restore_sigcontext()
364 if (v_regs != NULL && (msr & MSR_VEC) != 0) restore_sigcontext()
386 if ((msr & MSR_VSX) != 0) restore_sigcontext()
408 unsigned long msr; restore_tm_sigcontexts() local
429 err |= __get_user(msr, &sc->gp_regs[PT_MSR]); restore_tm_sigcontexts()
431 if (MSR_TM_RESV(msr)) restore_tm_sigcontexts()
435 regs->msr = (regs->msr & ~MSR_TS_MASK) | (msr & MSR_TS_MASK); restore_tm_sigcontexts()
438 regs->msr = (regs->msr & ~MSR_LE) | (msr & MSR_LE); restore_tm_sigcontexts()
474 regs->msr &= ~(MSR_FP | MSR_FE0 | MSR_FE1 | MSR_VEC | MSR_VSX); restore_tm_sigcontexts()
487 if (v_regs != NULL && tm_v_regs != NULL && (msr & MSR_VEC) != 0) { restore_tm_sigcontexts()
520 if (v_regs && ((msr & MSR_VSX) != 0)) { restore_tm_sigcontexts()
536 tm_recheckpoint(&current->thread, msr); restore_tm_sigcontexts()
539 if (msr & MSR_FP) { restore_tm_sigcontexts()
541 regs->msr |= (MSR_FP | current->thread.fpexc_mode); restore_tm_sigcontexts()
544 if (msr & MSR_VEC) { restore_tm_sigcontexts()
546 regs->msr |= MSR_VEC; restore_tm_sigcontexts()
669 unsigned long msr; sys_rt_sigreturn() local
682 if (__get_user(msr, &uc->uc_mcontext.gp_regs[PT_MSR])) sys_rt_sigreturn()
684 if (MSR_TM_ACTIVE(msr)) { sys_rt_sigreturn()
707 printk_ratelimited(regs->msr & MSR_64BIT ? fmt64 : fmt32, sys_rt_sigreturn()
735 if (MSR_TM_ACTIVE(regs->msr)) { handle_rt_signal64()
792 regs->msr &= ~MSR_LE; handle_rt_signal64()
793 regs->msr |= (MSR_KERNEL & MSR_LE); handle_rt_signal64()
811 printk_ratelimited(regs->msr & MSR_64BIT ? fmt64 : fmt32, handle_rt_signal64()
H A Dprocess.c87 MSR_TM_ACTIVE(tsk->thread.regs->msr) && giveup_fpu_maybe_transactional()
89 tsk->thread.tm_orig_msr = tsk->thread.regs->msr; giveup_fpu_maybe_transactional()
105 MSR_TM_ACTIVE(tsk->thread.regs->msr) && giveup_altivec_maybe_transactional()
107 tsk->thread.tm_orig_msr = tsk->thread.regs->msr; giveup_altivec_maybe_transactional()
129 * another process could get scheduled after the regs->msr flush_fp_to_thread()
136 if (tsk->thread.regs->msr & MSR_FP) { flush_fp_to_thread()
160 if (current->thread.regs && (current->thread.regs->msr & MSR_FP)) enable_kernel_fp()
176 if (current->thread.regs && (current->thread.regs->msr & MSR_VEC)) enable_kernel_altivec()
194 if (tsk->thread.regs->msr & MSR_VEC) { flush_altivec_to_thread()
212 if (current->thread.regs && (current->thread.regs->msr & MSR_VSX)) enable_kernel_vsx()
234 if (tsk->thread.regs->msr & MSR_VSX) { flush_vsx_to_thread()
253 if (current->thread.regs && (current->thread.regs->msr & MSR_SPE)) enable_kernel_spe()
267 if (tsk->thread.regs->msr & MSR_SPE) { flush_spe_to_thread()
543 msr_diff = thr->tm_orig_msr & ~thr->regs->msr; tm_reclaim_thread()
572 tm_reclaim(thr, thr->regs->msr, cause); tm_reclaim_thread()
577 * flush_fp_to_thread(), so update thr->regs->msr to tm_reclaim_thread()
580 thr->regs->msr |= msr_diff; tm_reclaim_thread()
606 if (!MSR_TM_ACTIVE(thr->regs->msr)) tm_reclaim_task()
615 thr->tm_orig_msr = thr->regs->msr; tm_reclaim_task()
618 "ccr=%lx, msr=%lx, trap=%lx)\n", tm_reclaim_task()
620 thr->regs->ccr, thr->regs->msr, tm_reclaim_task()
664 unsigned long msr; tm_recheckpoint_new_task() local
680 if (!MSR_TM_ACTIVE(new->thread.regs->msr)){ tm_recheckpoint_new_task()
684 msr = new->thread.tm_orig_msr; tm_recheckpoint_new_task()
687 "(new->msr 0x%lx, new->origmsr 0x%lx)\n", tm_recheckpoint_new_task()
688 new->pid, new->thread.regs->msr, msr); tm_recheckpoint_new_task()
691 tm_recheckpoint(&new->thread, msr); tm_recheckpoint_new_task()
694 if (msr & MSR_FP) { tm_recheckpoint_new_task()
696 new->thread.regs->msr |= tm_recheckpoint_new_task()
700 if (msr & MSR_VEC) { tm_recheckpoint_new_task()
702 new->thread.regs->msr |= MSR_VEC; tm_recheckpoint_new_task()
706 if (msr & MSR_VSX) tm_recheckpoint_new_task()
707 new->thread.regs->msr |= MSR_VSX; tm_recheckpoint_new_task()
710 "(kernel msr 0x%lx)\n", tm_recheckpoint_new_task()
741 if (!MSR_TM_ACTIVE(regs->msr)) restore_tm_state()
744 msr_diff = current->thread.tm_orig_msr & ~regs->msr; restore_tm_state()
749 regs->msr |= current->thread.fpexc_mode; restore_tm_state()
755 regs->msr |= msr_diff; restore_tm_state()
796 if (prev->thread.regs && (prev->thread.regs->msr & MSR_FP)) __switch_to()
810 if (prev->thread.regs && (prev->thread.regs->msr & MSR_VEC)) __switch_to()
814 if (prev->thread.regs && (prev->thread.regs->msr & MSR_VSX)) __switch_to()
826 if ((prev->thread.regs && (prev->thread.regs->msr & MSR_SPE))) __switch_to()
836 new->thread.regs->msr |= MSR_VEC; __switch_to()
840 new->thread.regs->msr |= MSR_VSX; __switch_to()
847 new->thread.regs->msr |= MSR_SPE; __switch_to()
935 if (!(regs->msr & MSR_IR)) show_instructions()
1021 printk("MSR: "REG" ", regs->msr); show_regs()
1022 printbits(regs->msr, msr_bits); show_regs()
1037 if (MSR_TM_ACTIVE(regs->msr)) show_regs()
1260 regs->msr = MSR_USER; start_thread()
1302 regs->msr = MSR_USER64; start_thread()
1306 regs->msr = MSR_USER32; start_thread()
1330 regs->msr |= MSR_TM; start_thread()
1384 if (regs != NULL && (regs->msr & MSR_FP) != 0) set_fpexc_mode()
1385 regs->msr = (regs->msr & ~(MSR_FE0|MSR_FE1)) set_fpexc_mode()
1433 regs->msr &= ~MSR_LE; set_endian()
1435 regs->msr |= MSR_LE; set_endian()
1454 if (regs->msr & MSR_LE) { get_endian()
H A Dsignal_32.c412 unsigned long msr = regs->msr; save_user_regs() local
430 msr |= MSR_VEC; save_user_regs()
432 /* else assert((regs->msr & MSR_VEC) == 0) */ save_user_regs()
452 msr &= ~MSR_VSX; save_user_regs()
464 msr |= MSR_VSX; save_user_regs()
476 msr |= MSR_SPE; save_user_regs()
478 /* else assert((regs->msr & MSR_SPE) == 0) */ save_user_regs()
485 if (__put_user(msr, &frame->mc_gregs[PT_MSR])) save_user_regs()
519 unsigned long msr = regs->msr; save_tm_user_regs() local
526 regs->msr &= ~MSR_TS_MASK; save_tm_user_regs()
542 if (__put_user((msr >> 32), &tm_frame->mc_gregs[PT_MSR])) save_tm_user_regs()
552 if (msr & MSR_VEC) { save_tm_user_regs()
567 msr |= MSR_VEC; save_tm_user_regs()
580 if (msr & MSR_VEC) { save_tm_user_regs()
593 if (msr & MSR_FP) { save_tm_user_regs()
612 if (msr & MSR_VSX) { save_tm_user_regs()
621 msr |= MSR_VSX; save_tm_user_regs()
635 msr |= MSR_SPE; save_tm_user_regs()
643 if (__put_user(msr, &frame->mc_gregs[PT_MSR])) save_tm_user_regs()
667 unsigned long msr; restore_user_regs() local
680 err |= __get_user(msr, &sr->mc_gregs[PT_MSR]); restore_user_regs()
688 regs->msr = (regs->msr & ~MSR_LE) | (msr & MSR_LE); restore_user_regs()
704 regs->msr &= ~MSR_VEC; restore_user_regs()
705 if (msr & MSR_VEC) { restore_user_regs()
728 regs->msr &= ~MSR_VSX; restore_user_regs()
729 if (msr & MSR_VSX) { restore_user_regs()
744 regs->msr &= ~(MSR_FP | MSR_FE0 | MSR_FE1); restore_user_regs()
749 regs->msr &= ~MSR_SPE; restore_user_regs()
750 if (msr & MSR_SPE) { restore_user_regs()
777 unsigned long msr, msr_hi; restore_tm_user_regs() local
794 err |= __get_user(msr, &sr->mc_gregs[PT_MSR]); restore_tm_user_regs()
799 regs->msr = (regs->msr & ~MSR_LE) | (msr & MSR_LE); restore_tm_user_regs()
811 regs->msr &= ~MSR_VEC; restore_tm_user_regs()
812 if (msr & MSR_VEC) { restore_tm_user_regs()
837 regs->msr &= ~(MSR_FP | MSR_FE0 | MSR_FE1); restore_tm_user_regs()
844 regs->msr &= ~MSR_VSX; restore_tm_user_regs()
845 if (msr & MSR_VSX) { restore_tm_user_regs()
864 regs->msr &= ~MSR_SPE; restore_tm_user_regs()
865 if (msr & MSR_SPE) { restore_tm_user_regs()
886 regs->msr = (regs->msr & ~MSR_TS_MASK) | (msr_hi & MSR_TS_MASK); restore_tm_user_regs()
895 tm_recheckpoint(&current->thread, msr); restore_tm_user_regs()
898 if (msr & MSR_FP) { restore_tm_user_regs()
900 regs->msr |= (MSR_FP | current->thread.fpexc_mode); restore_tm_user_regs()
903 if (msr & MSR_VEC) { restore_tm_user_regs()
905 regs->msr |= MSR_VEC; restore_tm_user_regs()
1026 if (MSR_TM_ACTIVE(regs->msr)) { handle_rt_signal32()
1061 regs->msr &= ~MSR_LE; handle_rt_signal32()
1062 regs->msr |= (MSR_KERNEL & MSR_LE); handle_rt_signal32()
1311 unsigned long new_msr = regs->msr; sys_debug_setcontext()
1361 regs->msr = new_msr; sys_debug_setcontext()
1451 if (MSR_TM_ACTIVE(regs->msr)) { handle_signal32()
1477 regs->msr &= ~MSR_LE; handle_signal32()
H A Dtraps.c255 printk_ratelimited(regs->msr & MSR_64BIT ? fmt64 : fmt32, _exception()
283 if (!(regs->msr & MSR_RI)) system_reset_exception()
292 * regs->nip and regs->msr contains srr0 and ssr1.
330 unsigned long msr = regs->msr; check_io_access() local
334 if (((msr & 0xffff0000) == 0 || (msr & (0x80000 | 0x40000))) check_io_access()
357 regs->msr |= MSR_RI; check_io_access()
387 #define get_reason(regs) ((regs)->msr)
388 #define get_mc_reason(regs) ((regs)->msr)
395 #define single_stepping(regs) ((regs)->msr & MSR_SE)
396 #define clear_single_step(regs) ((regs)->msr &= ~MSR_SE)
692 printk("Unknown values in msr\n"); machine_check_generic()
739 if (!(regs->msr & MSR_RI)) machine_check_exception()
770 regs->nip, regs->msr, regs->trap); unknown_exception()
912 if ((regs->msr & MSR_64BIT) == 0) emulate_string_inst()
990 if (MSR_TM_TRANSACTIONAL(regs->msr)) { tm_abort_check()
1159 if (!(regs->msr & MSR_PR) && /* not user-mode */ program_check_exception()
1194 "at %lx (msr 0x%x)\n", regs->nip, reason); program_check_exception()
1254 regs->msr |= REASON_ILLEGAL; emulation_assist_interrupt()
1309 regs->nip, regs->msr); nonrecoverable_exception()
1410 hv ? "Hypervisor " : "", facility, regs->nip, regs->msr); facility_unavailable_exception()
1428 regs->nip, regs->msr); fp_unavailable_tm()
1441 regs->msr |= (MSR_FP | current->thread.fpexc_mode); fp_unavailable_tm()
1452 if (regs->msr & MSR_VEC) { fp_unavailable_tm()
1455 regs->msr |= MSR_VSX; fp_unavailable_tm()
1467 regs->nip, regs->msr); altivec_unavailable_tm()
1469 regs->msr |= MSR_VEC; altivec_unavailable_tm()
1473 if (regs->msr & MSR_FP) { altivec_unavailable_tm()
1475 regs->msr |= MSR_VSX; altivec_unavailable_tm()
1481 unsigned long orig_msr = regs->msr; vsx_unavailable_tm()
1492 regs->nip, regs->msr); vsx_unavailable_tm()
1498 regs->msr |= MSR_VSX; vsx_unavailable_tm()
1505 regs->msr |= MSR_VEC | MSR_FP | current->thread.fpexc_mode | vsx_unavailable_tm()
1511 tm_recheckpoint(&current->thread, regs->msr & ~orig_msr); vsx_unavailable_tm()
1596 regs->msr |= MSR_DE; handle_debug()
1615 regs->msr &= ~MSR_DE; DebugException()
1626 regs->msr |= MSR_DE; DebugException()
1637 regs->msr &= ~MSR_DE; DebugException()
1656 regs->msr |= MSR_DE; DebugException()
1672 regs->nip, regs->msr, regs->trap, print_tainted()); TAUException()
1779 if (regs->msr & MSR_SPE) SPEFloatingPointRoundException()
H A Dptrace.c100 REG_OFFSET_NAME(msr),
154 * Set of msr bits that gdb can change on behalf of a process.
173 return task->thread.regs->msr | task->thread.fpexc_mode; get_user_msr()
176 static int set_user_msr(struct task_struct *task, unsigned long msr) set_user_msr() argument
178 task->thread.regs->msr &= ~MSR_DEBUGCHANGE; set_user_msr()
179 task->thread.regs->msr |= msr & MSR_DEBUGCHANGE; set_user_msr()
281 0, offsetof(struct pt_regs, msr)); gpr_get()
283 unsigned long msr = get_user_msr(target); gpr_get() local
284 ret = user_regset_copyout(&pos, &count, &kbuf, &ubuf, &msr, gpr_get()
285 offsetof(struct pt_regs, msr), gpr_get()
286 offsetof(struct pt_regs, msr) + gpr_get()
287 sizeof(msr)); gpr_get()
291 offsetof(struct pt_regs, msr) + sizeof(long)); gpr_get()
330 offsetof(struct pt_regs, msr) + sizeof(long)); gpr_set()
860 regs->msr |= MSR_DE; user_enable_single_step()
862 regs->msr &= ~MSR_BE; user_enable_single_step()
863 regs->msr |= MSR_SE; user_enable_single_step()
877 regs->msr |= MSR_DE; user_enable_block_step()
879 regs->msr &= ~MSR_SE; user_enable_block_step()
880 regs->msr |= MSR_BE; user_enable_block_step()
908 regs->msr &= ~MSR_DE; user_disable_single_step()
911 regs->msr &= ~(MSR_SE | MSR_BE); user_disable_single_step()
1032 task->thread.regs->msr &= ~MSR_DE; ptrace_set_debugreg()
1054 task->thread.regs->msr |= MSR_DE; ptrace_set_debugreg()
1154 child->thread.regs->msr |= MSR_DE; set_instruction_bp()
1269 child->thread.regs->msr |= MSR_DE; set_dac()
1363 child->thread.regs->msr |= MSR_DE; set_dac_range()
1496 child->thread.regs->msr &= ~MSR_DE; ppc_del_hwdebug()
H A Dkprobes.c131 kcb->kprobe_saved_msr = regs->msr; set_current_kprobe()
165 regs->msr &= ~MSR_SINGLESTEP; kprobe_handler()
166 regs->msr |= kcb->kprobe_saved_msr; kprobe_handler()
177 kcb->kprobe_saved_msr = regs->msr; kprobe_handler()
382 regs->msr |= kcb->kprobe_saved_msr; post_kprobe_handler()
394 * if somebody else is singlestepping across a probe point, msr post_kprobe_handler()
398 if (regs->msr & MSR_SINGLESTEP) post_kprobe_handler()
421 regs->msr &= ~MSR_SINGLESTEP; /* Turn off 'trace' bits */ kprobe_fault_handler()
422 regs->msr |= kcb->kprobe_saved_msr; kprobe_fault_handler()
H A Dppc32.h22 unsigned int msr; member in struct:pt_regs32
H A Dkgdb.c253 PACK64(ptr, regs->msr); sleeping_thread_to_gdb_regs()
341 { "msr", GDB_SIZEOF_REG, offsetof(struct pt_regs, msr) },
355 /* pc, msr, ls... registers 64 -> 69 */ dbg_get_reg()
381 /* pc, msr, ls... registers 64 -> 69 */ dbg_set_reg()
431 linux_regs->msr |= MSR_DE; kgdb_arch_handle_exception()
433 linux_regs->msr |= MSR_SE; kgdb_arch_handle_exception()
H A Dmce.c90 mce->srr1 = regs->msr; save_mce_event()
101 srr1 = regs->msr; save_mce_event()
H A Dmce_power.c274 srr1 = regs->msr; __machine_check_early_realmode_p7()
343 srr1 = regs->msr; __machine_check_early_realmode_p8()
H A Dsignal.c190 if (MSR_TM_ACTIVE(regs->msr)) { get_tm_stackpointer()
192 if (MSR_TM_TRANSACTIONAL(regs->msr)) get_tm_stackpointer()
H A Dhead_booke.h226 #define EXC_XFER_TEMPLATE(hdlr, trap, msr, copyee, tfer, ret) \
229 lis r10,msr@h; \
230 ori r10,r10,msr@l; \
/linux-4.1.27/arch/microblaze/include/asm/
H A Dkgdb.h12 * 6 32-bit special registers (pc, msr, ear, esr, fsr, btr)
H A Dsetup.h39 unsigned int fdt, unsigned int msr, unsigned int tlb0,
/linux-4.1.27/sound/pci/ctxfi/
H A Dctsrc.c186 if (src->rsc.msr > 1) { src_commit_write()
197 for (i = 1; i < src->rsc.msr; i++) { src_commit_write()
232 unsigned int rsr, msr; src_default_config_memrd() local
236 for (rsr = 0, msr = src->rsc.msr; msr > 1; msr >>= 1) src_default_config_memrd()
258 for (msr = 1; msr < src->rsc.msr; msr++) { src_default_config_memrd()
300 unsigned int rsr, msr; src_default_config_arcrw() local
305 for (rsr = 0, msr = src->rsc.msr; msr > 1; msr >>= 1) src_default_config_arcrw()
327 for (msr = 0; msr < src->rsc.msr; msr++) { src_default_config_arcrw()
369 err = rsc_init(&p->rsc, idx + i, SRC, desc->msr, mgr->mgr.hw); src_rsc_init()
496 for (i = 0; i < src->rsc.msr; i++) { src_enable_s()
512 for (i = 0; i < src->rsc.msr; i++) { src_enable()
528 for (i = 0; i < src->rsc.msr; i++) { src_disable()
631 for (i = 0; i < srcimp->rsc.msr; i++) { srcimp_map()
654 for (i = 0; i < srcimp->rsc.msr; i++) { srcimp_unmap()
677 SRCIMP, desc->msr, mgr->mgr.hw); srcimp_rsc_init()
682 srcimp->imappers = kzalloc(sizeof(struct imapper)*desc->msr, srcimp_rsc_init()
735 for (i = 0; i < desc->msr; i++) { get_srcimp_rsc()
773 for (i = 0; i < srcimp->rsc.msr; i++) put_srcimp_rsc()
H A Dctatc.c261 desc.msr = atc->msr; atc_pcm_playback_prepare()
268 (atc->rsr * atc->msr)); atc_pcm_playback_prepare()
283 mix_dsc.msr = atc->msr; atc_pcm_playback_prepare()
387 max_cisz = src->multi * src->rsc.msr; atc_pcm_playback_start()
448 max_cisz = src->multi * src->rsc.msr; atc_pcm_playback_position()
456 unsigned int msr:8; member in struct:src_node_conf_t
468 pitch = atc_get_pitch((atc->rsr * atc->msr), setup_src_node_conf()
472 if (1 == atc->msr) { /* FIXME: do we really need SRC here if pitch==1 */ setup_src_node_conf()
475 conf[0].mix_msr = conf[0].imp_msr = conf[0].msr = 1; setup_src_node_conf()
477 } else if (2 <= atc->msr) { setup_src_node_conf()
481 conf[0].pitch = (atc->msr << 24); setup_src_node_conf()
482 conf[0].msr = conf[0].mix_msr = 1; setup_src_node_conf()
483 conf[0].imp_msr = atc->msr; setup_src_node_conf()
487 conf[1].msr = conf[1].mix_msr = conf[1].imp_msr = 1; setup_src_node_conf()
494 conf[0].msr = conf[0].mix_msr setup_src_node_conf()
495 = conf[0].imp_msr = atc->msr; setup_src_node_conf()
529 pitch = atc_get_pitch((atc->rsr * atc->msr), atc_pcm_capture_get_resources()
539 n_amixer += multi * atc->msr; atc_pcm_capture_get_resources()
540 n_srcimp += multi * atc->msr; atc_pcm_capture_get_resources()
567 src_dsc.msr = src_node_conf[i/multi].msr; atc_pcm_capture_get_resources()
585 mix_dsc.msr = atc->msr; atc_pcm_capture_get_resources()
587 mix_dsc.msr = src_node_conf[(i-n_sum*2)/multi].mix_msr; atc_pcm_capture_get_resources()
589 mix_dsc.msr = 1; atc_pcm_capture_get_resources()
600 sum_dsc.msr = atc->msr; atc_pcm_capture_get_resources()
605 pitch = atc_get_pitch((atc->rsr * atc->msr), atc_pcm_capture_get_resources()
610 srcimp_dsc.msr = src_node_conf[i/multi].imp_msr; atc_pcm_capture_get_resources()
612 srcimp_dsc.msr = (pitch <= 0x8000000) ? atc->msr : 1; atc_pcm_capture_get_resources()
614 srcimp_dsc.msr = 1; atc_pcm_capture_get_resources()
626 src_dsc.msr = 1; atc_pcm_capture_get_resources()
692 pitch = atc_get_pitch((atc->rsr * atc->msr), atc_pcm_capture_prepare()
700 for (j = 0; j < atc->msr; j++) { atc_pcm_capture_prepare()
802 desc.msr = 1; spdif_passthru_playback_get_resources()
803 while (apcm->substream->runtime->rate > (rsr * desc.msr)) spdif_passthru_playback_get_resources()
804 desc.msr <<= 1; spdif_passthru_playback_get_resources()
811 pitch = atc_get_pitch(apcm->substream->runtime->rate, (rsr * desc.msr)); spdif_passthru_playback_get_resources()
827 mix_dsc.msr = desc.msr; spdif_passthru_playback_get_resources()
1124 da_dsc.msr = state ? 1 : atc->msr; atc_spdif_out_passthru()
1345 info.msr = atc->msr; atc_create_hw_devs()
1398 da_desc.msr = atc->msr; atc_get_resources()
1415 src_dsc.msr = atc->msr; atc_get_resources()
1427 srcimp_dsc.msr = 8; atc_get_resources()
1438 sum_dsc.msr = atc->msr; atc_get_resources()
1577 info.msr = atc->msr; atc_hw_resume()
1678 unsigned int rsr, unsigned int msr, ct_atc_create()
1700 atc->msr = msr; ct_atc_create()
1677 ct_atc_create(struct snd_card *card, struct pci_dev *pci, unsigned int rsr, unsigned int msr, int chip_type, unsigned int ssid, struct ct_atc **ratc) ct_atc_create() argument
H A Dctdaio.c166 entry = kzalloc((sizeof(*entry) * daio->rscl.msr), GFP_KERNEL); dao_set_left_input()
174 for (i = 0; i < daio->rscl.msr; i++, entry++) { dao_set_left_input()
195 entry = kzalloc((sizeof(*entry) * daio->rscr.msr), GFP_KERNEL); dao_set_right_input()
203 for (i = 0; i < daio->rscr.msr; i++, entry++) { dao_set_right_input()
207 dao->imappers[daio->rscl.msr + i] = entry; dao_set_right_input()
230 for (i = 1; i < daio->rscl.msr; i++) { dao_clear_left_input()
248 if (!dao->imappers[daio->rscl.msr]) dao_clear_right_input()
251 entry = dao->imappers[daio->rscl.msr]; dao_clear_right_input()
254 for (i = 1; i < daio->rscr.msr; i++) { dao_clear_right_input()
255 entry = dao->imappers[daio->rscl.msr + i]; dao_clear_right_input()
257 dao->imappers[daio->rscl.msr + i] = NULL; dao_clear_right_input()
260 kfree(dao->imappers[daio->rscl.msr]); dao_clear_right_input()
261 dao->imappers[daio->rscl.msr] = NULL; dao_clear_right_input()
291 static int dai_set_srt_msr(struct dai *dai, unsigned int msr) dai_set_srt_msr() argument
295 for (rsr = 0; msr > 1; msr >>= 1) dai_set_srt_msr()
349 err = rsc_init(&daio->rscl, idx_l, DAIO, desc->msr, hw); daio_rsc_init()
353 err = rsc_init(&daio->rscr, idx_r, DAIO, desc->msr, hw); daio_rsc_init()
401 dao->imappers = kzalloc(sizeof(void *)*desc->msr*2, GFP_KERNEL); dao_rsc_init()
417 conf = (desc->msr & 0x7) | (desc->passthru << 3); dao_rsc_init()
440 if (dao->imappers[dao->daio.rscl.msr]) dao_rsc_uninit()
459 dsc.msr = desc->msr; dao_rsc_reinit()
471 unsigned int rsr, msr; dai_rsc_init() local
483 for (rsr = 0, msr = desc->msr; msr > 1; msr >>= 1) dai_rsc_init()
H A Dctdaio.h69 unsigned int msr:4; member in struct:dao_desc
87 int (*set_srt_msr)(struct dai *dai, unsigned int msr);
96 unsigned int msr:4; member in struct:daio_desc
H A Dctresource.h39 u32 msr:4; /* The Master Sample Rate a resource working on */ member in struct:rsc
54 rsc_init(struct rsc *rsc, u32 idx, enum RSCTYP type, u32 msr, struct hw *hw);
H A Dctamixer.c133 for (i = 0; i < amixer->rsc.msr; i++) { amixer_commit_write()
207 AMIXER, desc->msr, mgr->mgr.hw); amixer_rsc_init()
252 for (i = 0; i < desc->msr; i++) { get_amixer_rsc()
290 for (i = 0; i < amixer->rsc.msr; i++) put_amixer_rsc()
373 err = rsc_init(&sum->rsc, sum->idx[0], SUM, desc->msr, mgr->mgr.hw); sum_rsc_init()
407 for (i = 0; i < desc->msr; i++) { get_sum_rsc()
445 for (i = 0; i < sum->rsc.msr; i++) put_sum_rsc()
H A Dctresource.c119 for (i = 0; (i < 8) && (!(rsc->msr & (0x1 << i))); ) rsc_next_conj()
138 rsc_init(struct rsc *rsc, u32 idx, enum RSCTYP type, u32 msr, struct hw *hw) rsc_init() argument
145 rsc->msr = msr; rsc_init()
204 rsc->msr = 0; rsc_uninit()
H A Dctamixer.h34 unsigned int msr; member in struct:sum_desc
79 unsigned int msr; member in struct:amixer_desc
H A Dcthw20k2.c1139 unsigned int msr; /* master sample rate in rsrs */ member in struct:dac_conf
1143 unsigned int msr; /* master sample rate in rsrs */ member in struct:adc_conf
1149 unsigned int msr; /* master sample rate in rsrs */ member in struct:daio_conf
1163 if (1 == info->msr) { hw_daio_init()
1167 } else if (2 == info->msr) { hw_daio_init()
1185 } else if ((4 == info->msr) && (hw->model == CTSB1270)) { hw_daio_init()
1227 if (2 == info->msr) { hw_daio_init()
1230 } else if (4 == info->msr) { hw_daio_init()
1641 if (1 == info->msr) hw_dac_init()
1643 else if (2 == info->msr) hw_dac_init()
1731 if (1 == info->msr) { hw_dac_init()
1736 } else if (2 == info->msr) { hw_dac_init()
1864 if (1 == info->msr) hw_adc_init()
1866 else if (2 == info->msr) hw_adc_init()
1886 if (1 == info->msr) { hw_adc_init()
1890 } else if ((2 == info->msr) || (4 == info->msr)) { hw_adc_init()
1896 "Invalid master sampling rate (msr %d)!!!\n", hw_adc_init()
1897 info->msr); hw_adc_init()
2182 daio_info.msr = info->msr; hw_card_init()
2187 dac_info.msr = info->msr; hw_card_init()
2192 adc_info.msr = info->msr; hw_card_init()
H A Dctatc.h81 unsigned int msr; /* master sample rate in rsr */ member in struct:ct_atc
156 unsigned int rsr, unsigned int msr, int chip_type,
H A Dctsrc.h82 unsigned char msr; member in struct:src_desc
123 unsigned int msr; member in struct:srcimp_desc
/linux-4.1.27/arch/powerpc/kvm/
H A Dtrace_booke.h46 __field( unsigned long, msr )
55 __entry->msr = vcpu->arch.shared->msr;
61 " | msr=0x%lx"
67 __entry->msr,
H A Dbook3s_pr.c55 ulong msr);
67 ulong msr = kvmppc_get_msr(vcpu); kvmppc_is_split_real() local
68 return (msr & (MSR_IR|MSR_DR)) == MSR_DR; kvmppc_is_split_real()
73 ulong msr = kvmppc_get_msr(vcpu); kvmppc_fixup_split_real() local
77 if ((msr & (MSR_IR|MSR_DR)) != MSR_DR) kvmppc_fixup_split_real()
337 static void kvmppc_set_msr_pr(struct kvm_vcpu *vcpu, u64 msr) kvmppc_set_msr_pr() argument
342 printk(KERN_INFO "KVM: Set MSR to 0x%llx\n", msr); kvmppc_set_msr_pr()
345 msr &= to_book3s(vcpu)->msr_mask; kvmppc_set_msr_pr()
346 kvmppc_set_msr_fast(vcpu, msr); kvmppc_set_msr_pr()
349 if (msr & MSR_POW) { kvmppc_set_msr_pr()
356 msr &= ~MSR_POW; kvmppc_set_msr_pr()
357 kvmppc_set_msr_fast(vcpu, msr); kvmppc_set_msr_pr()
372 if (!(msr & MSR_PR) && vcpu->arch.magic_page_pa) { kvmppc_set_msr_pr()
375 if (msr & MSR_DR) kvmppc_set_msr_pr()
391 !(old_msr & MSR_PR) && !(old_msr & MSR_SF) && (msr & MSR_SF)) { kvmppc_set_msr_pr()
598 u64 msr = kvmppc_get_msr(vcpu); kvmppc_handle_pagefault() local
601 kvmppc_set_msr_fast(vcpu, msr | (ssrr1 & 0xf8000000ULL)); kvmppc_handle_pagefault()
607 u64 msr = kvmppc_get_msr(vcpu); kvmppc_handle_pagefault() local
611 kvmppc_set_msr_fast(vcpu, msr | (ssrr1 & 0xf8000000ULL)); kvmppc_handle_pagefault()
648 void kvmppc_giveup_ext(struct kvm_vcpu *vcpu, ulong msr) kvmppc_giveup_ext() argument
656 if (msr & MSR_VSX) kvmppc_giveup_ext()
657 msr |= MSR_FP | MSR_VEC; kvmppc_giveup_ext()
659 msr &= vcpu->arch.guest_owned_ext; kvmppc_giveup_ext()
660 if (!msr) kvmppc_giveup_ext()
664 printk(KERN_INFO "Giving up ext 0x%lx\n", msr); kvmppc_giveup_ext()
667 if (msr & MSR_FP) { kvmppc_giveup_ext()
673 if (t->regs->msr & MSR_FP) kvmppc_giveup_ext()
679 if (msr & MSR_VEC) { kvmppc_giveup_ext()
680 if (current->thread.regs->msr & MSR_VEC) kvmppc_giveup_ext()
686 vcpu->arch.guest_owned_ext &= ~(msr | MSR_VSX); kvmppc_giveup_ext()
711 ulong msr) kvmppc_handle_ext()
719 if (!(kvmppc_get_msr(vcpu) & msr)) { kvmppc_handle_ext()
724 if (msr == MSR_VSX) { kvmppc_handle_ext()
738 msr = MSR_FP | MSR_VEC | MSR_VSX; kvmppc_handle_ext()
742 msr &= ~vcpu->arch.guest_owned_ext; kvmppc_handle_ext()
743 if (!msr) kvmppc_handle_ext()
747 printk(KERN_INFO "Loading up ext 0x%lx\n", msr); kvmppc_handle_ext()
750 if (msr & MSR_FP) { kvmppc_handle_ext()
758 if (msr & MSR_VEC) { kvmppc_handle_ext()
768 t->regs->msr |= msr; kvmppc_handle_ext()
769 vcpu->arch.guest_owned_ext |= msr; kvmppc_handle_ext()
783 lost_ext = vcpu->arch.guest_owned_ext & ~current->thread.regs->msr; kvmppc_handle_lost_ext()
801 current->thread.regs->msr |= lost_ext; kvmppc_handle_lost_ext()
940 u64 msr = kvmppc_get_msr(vcpu); kvmppc_handle_exit_pr() local
941 msr |= shadow_srr1 & 0x58000000; kvmppc_handle_exit_pr()
942 kvmppc_set_msr_fast(vcpu, msr); kvmppc_handle_exit_pr()
1214 printk(KERN_EMERG "exit_nr=0x%x | pc=0x%lx | msr=0x%lx\n", kvmppc_handle_exit_pr()
1490 if (current->thread.regs->msr & MSR_FP) kvmppc_vcpu_run_pr()
1495 if (current->thread.regs->msr & MSR_VEC) kvmppc_vcpu_run_pr()
1501 if (current->thread.regs->msr & MSR_VSX) kvmppc_vcpu_run_pr()
710 kvmppc_handle_ext(struct kvm_vcpu *vcpu, unsigned int exit_nr, ulong msr) kvmppc_handle_ext() argument
H A Dbooke.c78 printk("pc: %08lx msr: %08llx\n", vcpu->arch.pc, vcpu->arch.shared->msr); kvmppc_dump_vcpu()
115 if (vcpu->arch.shared->msr & MSR_SPE) { kvmppc_vcpu_sync_spe()
140 if (!(current->thread.regs->msr & MSR_FP)) { kvmppc_load_guest_fp()
144 current->thread.regs->msr |= MSR_FP; kvmppc_load_guest_fp()
156 if (current->thread.regs->msr & MSR_FP) kvmppc_save_guest_fp()
168 vcpu->arch.shadow_msr |= vcpu->arch.shared->msr & MSR_FP; kvmppc_vcpu_sync_fpu()
181 if (!(current->thread.regs->msr & MSR_VEC)) { kvmppc_load_guest_altivec()
185 current->thread.regs->msr |= MSR_VEC; kvmppc_load_guest_altivec()
199 if (current->thread.regs->msr & MSR_VEC) kvmppc_save_guest_altivec()
211 vcpu->arch.shadow_msr |= vcpu->arch.shared->msr & MSR_DE; kvmppc_vcpu_sync_debug()
221 vcpu->arch.shared->msr |= MSR_DE; kvmppc_vcpu_sync_debug()
224 vcpu->arch.shared->msr &= ~MSR_DE; kvmppc_vcpu_sync_debug()
235 u32 old_msr = vcpu->arch.shared->msr; kvmppc_set_msr()
241 vcpu->arch.shared->msr = new_msr; kvmppc_set_msr()
389 ulong new_msr = vcpu->arch.shared->msr; kvmppc_booke_irqprio_deliver()
392 if (!(vcpu->arch.shared->msr & MSR_SF)) { kvmppc_booke_irqprio_deliver()
400 crit = crit && !(vcpu->arch.shared->msr & MSR_PR); kvmppc_booke_irqprio_deliver()
440 allowed = vcpu->arch.shared->msr & MSR_CE; kvmppc_booke_irqprio_deliver()
446 allowed = vcpu->arch.shared->msr & MSR_ME; kvmppc_booke_irqprio_deliver()
456 allowed = vcpu->arch.shared->msr & MSR_EE; kvmppc_booke_irqprio_deliver()
462 allowed = vcpu->arch.shared->msr & MSR_DE; kvmppc_booke_irqprio_deliver()
477 vcpu->arch.shared->msr); kvmppc_booke_irqprio_deliver()
481 vcpu->arch.shared->msr); kvmppc_booke_irqprio_deliver()
485 vcpu->arch.shared->msr); kvmppc_booke_irqprio_deliver()
489 vcpu->arch.shared->msr); kvmppc_booke_irqprio_deliver()
684 if (vcpu->arch.shared->msr & MSR_WE) { kvmppc_core_prepare_to_enter()
843 if (dbsr && (vcpu->arch.shared->msr & MSR_DE) && kvmppc_handle_debug()
848 if ((dbsr & DBSR_TIE) && !(vcpu->arch.shared->msr & MSR_DE)) kvmppc_handle_debug()
880 ulong r1, ip, msr, lr; kvmppc_fill_pt_regs() local
884 asm("mfmsr %0" : "=r"(msr)); kvmppc_fill_pt_regs()
890 regs->msr = msr; kvmppc_fill_pt_regs()
1092 if (vcpu->arch.shared->msr & (MSR_PR | MSR_GS)) { kvmppc_handle_exit()
1118 if (vcpu->arch.shared->msr & MSR_SPE) kvmppc_handle_exit()
1196 if (!(vcpu->arch.shared->msr & MSR_PR)) { kvmppc_handle_exit()
1210 if (!(vcpu->arch.shared->msr & MSR_PR) && kvmppc_handle_exit()
1231 if (!(vcpu->arch.shared->msr & MSR_PR) && kvmppc_handle_exit()
1383 vcpu->arch.shared->msr = 0; kvm_arch_vcpu_setup()
1428 regs->msr = vcpu->arch.shared->msr; kvm_arch_vcpu_ioctl_get_regs()
1456 kvmppc_set_msr(vcpu, regs->msr); kvm_arch_vcpu_ioctl_set_regs()
1938 if (!(vcpu->arch.shared->msr & MSR_PR) && kvmppc_xlate()
H A Dtrace_pr.h225 __field( unsigned long, msr )
235 __entry->msr = kvmppc_get_msr(vcpu);
242 " | msr=0x%lx"
249 __entry->msr,
H A Dbooke_emulate.c91 kvmppc_set_gpr(vcpu, rt, vcpu->arch.shared->msr); kvmppc_booke_emulate_op()
101 vcpu->arch.shared->msr = (vcpu->arch.shared->msr & ~MSR_EE) kvmppc_booke_emulate_op()
107 vcpu->arch.shared->msr = (vcpu->arch.shared->msr & ~MSR_EE) kvmppc_booke_emulate_op()
H A Dtrace_hv.h254 __field(unsigned long, msr)
263 __entry->msr = vcpu->arch.shregs.msr;
266 TP_printk("VCPU %d: trap=%s pc=0x%lx msr=0x%lx, ceded=%d",
269 __entry->pc, __entry->msr, __entry->ceded
H A Dbook3s.c236 u64 msr = kvmppc_get_msr(vcpu); kvmppc_core_queue_inst_storage() local
237 msr &= ~(SRR1_ISI_NOPT | SRR1_ISI_N_OR_G | SRR1_ISI_PROT); kvmppc_core_queue_inst_storage()
238 msr |= flags & (SRR1_ISI_NOPT | SRR1_ISI_N_OR_G | SRR1_ISI_PROT); kvmppc_core_queue_inst_storage()
239 kvmppc_set_msr_fast(vcpu, msr); kvmppc_core_queue_inst_storage()
475 regs->msr = kvmppc_get_msr(vcpu); kvm_arch_vcpu_ioctl_get_regs()
503 kvmppc_set_msr(vcpu, regs->msr); kvm_arch_vcpu_ioctl_set_regs()
691 void kvmppc_set_msr(struct kvm_vcpu *vcpu, u64 msr) kvmppc_set_msr() argument
693 vcpu->kvm->arch.kvm_ops->set_msr(vcpu, msr); kvmppc_set_msr()
H A Dbook3s_32_mmu.c369 u64 msr = kvmppc_get_msr(vcpu); kvmppc_mmu_book3s_32_esid_to_vsid() local
371 if (msr & (MSR_DR|MSR_IR)) { kvmppc_mmu_book3s_32_esid_to_vsid()
380 switch (msr & (MSR_DR|MSR_IR)) { kvmppc_mmu_book3s_32_esid_to_vsid()
400 if (msr & MSR_PR) kvmppc_mmu_book3s_32_esid_to_vsid()
H A De500.h218 return !!(vcpu->arch.shared->msr & (MSR_IS | MSR_DS)); get_cur_as()
223 return !!(vcpu->arch.shared->msr & MSR_PR); get_cur_pr()
266 if (get_tlb_ts(tlbe) != !!(vcpu->arch.shared->msr & MSR_IS)) tlbe_is_host_safe()
H A De500_mmu.c415 if (!(vcpu->arch.shared->msr & MSR_CM)) kvmppc_e500_emul_tlbwe()
500 unsigned int as = !!(vcpu->arch.shared->msr & MSR_IS); kvmppc_mmu_itlb_index()
507 unsigned int as = !!(vcpu->arch.shared->msr & MSR_DS); kvmppc_mmu_dtlb_index()
514 unsigned int as = !!(vcpu->arch.shared->msr & MSR_IS); kvmppc_mmu_itlb_miss()
521 unsigned int as = !!(vcpu->arch.shared->msr & MSR_DS); kvmppc_mmu_dtlb_miss()
H A Dbook3s_64_mmu.c585 u64 msr = kvmppc_get_msr(vcpu); kvmppc_mmu_book3s_64_esid_to_vsid() local
587 if (msr & (MSR_DR|MSR_IR)) { kvmppc_mmu_book3s_64_esid_to_vsid()
600 switch (msr & (MSR_DR|MSR_IR)) { kvmppc_mmu_book3s_64_esid_to_vsid()
/linux-4.1.27/tools/power/cpupower/utils/idle_monitor/
H A Dnhm_idle.c75 int msr; nhm_get_count() local
79 msr = MSR_CORE_C3_RESIDENCY; nhm_get_count()
82 msr = MSR_CORE_C6_RESIDENCY; nhm_get_count()
85 msr = MSR_PKG_C3_RESIDENCY; nhm_get_count()
88 msr = MSR_PKG_C6_RESIDENCY; nhm_get_count()
91 msr = MSR_TSC; nhm_get_count()
96 if (read_msr(cpu, msr, val)) nhm_get_count()
H A Dhsw_ext_idle.c67 int msr; hsw_ext_get_count() local
71 msr = MSR_PKG_C8_RESIDENCY; hsw_ext_get_count()
74 msr = MSR_PKG_C9_RESIDENCY; hsw_ext_get_count()
77 msr = MSR_PKG_C10_RESIDENCY; hsw_ext_get_count()
80 msr = MSR_TSC; hsw_ext_get_count()
85 if (read_msr(cpu, msr, val)) hsw_ext_get_count()
H A Dsnb_idle.c64 int msr; snb_get_count() local
68 msr = MSR_CORE_C7_RESIDENCY; snb_get_count()
71 msr = MSR_PKG_C2_RESIDENCY; snb_get_count()
74 msr = MSR_PKG_C7_RESIDENCY; snb_get_count()
77 msr = MSR_TSC; snb_get_count()
82 if (read_msr(cpu, msr, val)) snb_get_count()
/linux-4.1.27/arch/arm/kvm/
H A Dinterrupts_head.S104 msr SP_\mode, r2
105 msr LR_\mode, r3
106 msr SPSR_\mode, r4
115 msr r8_fiq, r2
116 msr r9_fiq, r3
117 msr r10_fiq, r4
118 msr r11_fiq, r5
119 msr r12_fiq, r6
120 msr SP_fiq, r7
121 msr LR_fiq, r8
122 msr SPSR_fiq, r9
130 msr SP_usr, r2
135 msr ELR_hyp, r2
149 msr SP_\mode, r2
150 msr LR_\mode, r3
151 msr SPSR_\mode, r4
169 msr r8_fiq, r2
170 msr r9_fiq, r3
171 msr r10_fiq, r4
172 msr r11_fiq, r5
173 msr r12_fiq, r6
174 msr SP_fiq, r7
175 msr LR_fiq, r8
176 msr SPSR_fiq, r9
181 msr ELR_hyp, r2
182 msr SPSR_cxsf, r3
187 msr SP_usr, r2
/linux-4.1.27/arch/powerpc/lib/
H A Dsstep.c51 static unsigned long truncate_if_32bit(unsigned long msr, unsigned long val) truncate_if_32bit() argument
54 if ((msr & MSR_64BIT) == 0) truncate_if_32bit()
104 return truncate_if_32bit(regs->msr, ea); dform_ea()
121 return truncate_if_32bit(regs->msr, ea); dsform_ea()
140 return truncate_if_32bit(regs->msr, ea); xform_ea()
539 if (!(regs->msr & MSR_64BIT)) set_cr0()
560 if (!(regs->msr & MSR_64BIT)) { add_with_carry()
661 regs->nip = truncate_if_32bit(regs->msr, regs->nip); analyse_instr()
665 regs->nip = truncate_if_32bit(regs->msr, imm); analyse_instr()
683 regs->link = truncate_if_32bit(regs->msr, regs->nip + 4); analyse_instr()
684 imm = truncate_if_32bit(regs->msr, imm); analyse_instr()
700 regs->nip = truncate_if_32bit(regs->msr, regs->nip + 4); analyse_instr()
701 imm = truncate_if_32bit(regs->msr, imm); analyse_instr()
709 if (regs->msr & MSR_PR) analyse_instr()
944 if (regs->msr & MSR_PR) analyse_instr()
950 if (regs->msr & MSR_PR) analyse_instr()
958 if (regs->msr & MSR_PR) analyse_instr()
1378 if (!(regs->msr & MSR_VEC)) analyse_instr()
1385 if (!(regs->msr & MSR_VEC)) analyse_instr()
1450 op->ea = truncate_if_32bit(regs->msr, analyse_instr()
1457 if (!(regs->msr & MSR_FP)) analyse_instr()
1464 if (!(regs->msr & MSR_FP)) analyse_instr()
1471 if (!(regs->msr & MSR_FP)) analyse_instr()
1478 if (!(regs->msr & MSR_FP)) analyse_instr()
1506 op->ea = truncate_if_32bit(regs->msr, analyse_instr()
1522 if (!(regs->msr & MSR_VSX)) analyse_instr()
1530 if (!(regs->msr & MSR_VSX)) analyse_instr()
1597 if (!(regs->msr & MSR_FP)) analyse_instr()
1605 if (!(regs->msr & MSR_FP)) analyse_instr()
1613 if (!(regs->msr & MSR_FP)) analyse_instr()
1621 if (!(regs->msr & MSR_FP)) analyse_instr()
1670 regs->nip = truncate_if_32bit(regs->msr, regs->nip + 4); analyse_instr()
1809 if (regs->msr & MSR_LE) emulate_step()
1832 if (regs->msr & MSR_LE) emulate_step()
1857 if (regs->msr & MSR_LE) emulate_step()
1870 if (regs->msr & MSR_LE) emulate_step()
1880 if (regs->msr & MSR_LE) emulate_step()
1887 if (regs->msr & MSR_LE) emulate_step()
1893 if (regs->msr & MSR_LE) emulate_step()
1911 if (regs->msr & MSR_LE) emulate_step()
1915 !(regs->msr & MSR_PR) && emulate_step()
1925 if (regs->msr & MSR_LE) emulate_step()
1935 if (regs->msr & MSR_LE) emulate_step()
1942 if (regs->msr & MSR_LE) emulate_step()
1948 if (regs->msr & MSR_LE) emulate_step()
1967 regs->gpr[op.reg] = regs->msr & MSR_MASK; emulate_step()
1976 regs->msr = (regs->msr & ~op.val) | (val & op.val); emulate_step()
1988 regs->msr ^= MSR_LE; emulate_step()
1994 regs->gpr[12] = regs->msr & MSR_MASK; emulate_step()
1997 regs->msr = MSR_KERNEL; emulate_step()
2013 regs->nip = truncate_if_32bit(regs->msr, regs->nip + 4); emulate_step()
/linux-4.1.27/arch/powerpc/platforms/pasemi/
H A Didle.c52 if (regs->msr & SRR1_WAKEMASK) pasemi_system_reset_exception()
55 switch (regs->msr & SRR1_WAKEMASK) { pasemi_system_reset_exception()
71 regs->msr |= MSR_RI; pasemi_system_reset_exception()
/linux-4.1.27/samples/kprobes/
H A Dkprobe_example.c32 " msr = 0x%lx\n", handler_pre()
33 p->addr, regs->nip, regs->msr); handler_pre()
59 printk(KERN_INFO "post_handler: p->addr = 0x%p, msr = 0x%lx\n", handler_post()
60 p->addr, regs->msr); handler_post()
/linux-4.1.27/drivers/cpufreq/
H A Dacpi-cpufreq.c45 #include <asm/msr.h>
83 static struct msr __percpu *msrs;
88 u64 msr; boost_state() local
93 msr = lo | ((u64)hi << 32); boost_state()
94 return !(msr & MSR_IA32_MISC_ENABLE_TURBO_DISABLE); boost_state()
97 msr = lo | ((u64)hi << 32); boost_state()
98 return !(msr & MSR_K7_HWCR_CPB_DIS); boost_state()
125 struct msr *reg = per_cpu_ptr(msrs, cpu); for_each_cpu()
214 static unsigned extract_msr(u32 msr, struct acpi_cpufreq_data *data) extract_msr() argument
220 msr &= AMD_MSR_RANGE; extract_msr()
222 msr &= INTEL_MSR_RANGE; extract_msr()
227 if (msr == perf->states[pos->driver_data].status) extract_msr()
258 struct msr_addr msr; member in union:drv_cmd::__anon3767
273 rdmsr(cmd->addr.msr.reg, cmd->val, h); do_drv_read()
293 rdmsr(cmd->addr.msr.reg, lo, hi); do_drv_write()
295 wrmsr(cmd->addr.msr.reg, lo, hi); do_drv_write()
298 wrmsr(cmd->addr.msr.reg, cmd->val, 0); do_drv_write()
341 cmd.addr.msr.reg = MSR_IA32_PERF_CTL; get_cur_val()
345 cmd.addr.msr.reg = MSR_AMD_PERF_CTL; get_cur_val()
439 cmd.addr.msr.reg = MSR_IA32_PERF_CTL; acpi_cpufreq_target()
444 cmd.addr.msr.reg = MSR_AMD_PERF_CTL; acpi_cpufreq_target()
H A Dspeedstep-centrino.c25 #include <asm/msr.h>
285 static unsigned extract_clock(unsigned msr, unsigned int cpu, int failsafe) extract_clock() argument
297 msr = (msr >> 8) & 0xff; extract_clock()
298 return msr * 100000; extract_clock()
305 msr &= 0xffff; extract_clock()
310 if (msr == per_cpu(centrino_model, cpu)->op_points[i].driver_data) extract_clock()
423 unsigned int msr, oldmsr = 0, h = 0, cpu = policy->cpu; centrino_target() local
462 msr = op_points->driver_data; centrino_target()
466 if (msr == (oldmsr & 0xffff)) { centrino_target()
467 pr_debug("no change needed - msr was and needs " centrino_target()
476 msr &= 0xffff; centrino_target()
477 oldmsr |= msr; centrino_target()
H A Damd_freq_sensitivity.c21 #include <asm/msr.h>
46 struct msr actual, reference; amd_powersave_bias_target()
H A Dsc520_freq.c26 #include <asm/msr.h>
H A Dsfi-cpufreq.c25 #include <asm/msr.h>
H A Delanfreq.c27 #include <asm/msr.h>
/linux-4.1.27/tools/power/x86/x86_energy_perf_policy/
H A Dx86_energy_perf_policy.c189 unsigned long long msr; get_msr() local
194 sprintf(msr_path, "/dev/cpu/%d/msr", cpu); get_msr()
197 printf("Try \"# modprobe msr\"\n"); get_msr()
202 retval = pread(fd, &msr, sizeof msr, offset); get_msr()
204 if (retval != sizeof msr) { get_msr()
209 return msr; get_msr()
219 sprintf(msr_path, "/dev/cpu/%d/msr", cpu); put_msr()
/linux-4.1.27/drivers/video/fbdev/geode/
H A Dsuspend_gx.c12 #include <asm/msr.h>
30 rdmsrl(MSR_GX_MSR_PADSEL, par->msr.padsel); gx_save_regs()
31 rdmsrl(MSR_GLCP_DOTPLL, par->msr.dotpll); gx_save_regs()
142 wrmsrl(MSR_GX_MSR_PADSEL, par->msr.padsel); gx_restore_video_proc()
179 gx_set_dotpll((uint32_t) (par->msr.dotpll >> 32)); gx_restore_regs()
H A Dlxfb_ops.c600 rdmsrl(MSR_LX_MSR_PADSEL, par->msr.padsel); lx_save_regs()
601 rdmsrl(MSR_GLCP_DOTPLL, par->msr.dotpll); lx_save_regs()
602 rdmsrl(MSR_LX_GLD_MSR_CONFIG, par->msr.dfglcfg); lx_save_regs()
603 rdmsrl(MSR_LX_SPARE_MSR, par->msr.dcspare); lx_save_regs()
673 wrmsrl(MSR_LX_SPARE_MSR, par->msr.dcspare); lx_restore_display_ctlr()
737 wrmsrl(MSR_LX_GLD_MSR_CONFIG, par->msr.dfglcfg); lx_restore_video_proc()
738 wrmsrl(MSR_LX_MSR_PADSEL, par->msr.padsel); lx_restore_video_proc()
772 lx_set_dotpll((u32) (par->msr.dotpll >> 32)); lx_restore_regs()
/linux-4.1.27/arch/powerpc/platforms/52xx/
H A Dmpc52xx_pm.c115 u32 msr, hid0; mpc52xx_pm_enter() local
143 msr = mfmsr(); mpc52xx_pm_enter()
144 mtmsr(msr & ~MSR_POW); mpc52xx_pm_enter()
163 mtmsr(msr & ~MSR_POW); mpc52xx_pm_enter()
165 mtmsr(msr); mpc52xx_pm_enter()
/linux-4.1.27/drivers/mfd/
H A Dezx-pcap.c46 u32 msr; member in struct:pcap_chip
151 pcap->msr |= 1 << irq_to_pcap(pcap, d->irq); pcap_mask_irq()
159 pcap->msr &= ~(1 << irq_to_pcap(pcap, d->irq)); pcap_unmask_irq()
174 ezx_pcap_write(pcap, PCAP_REG_MSR, pcap->msr); pcap_msr_work()
181 u32 msr, isr, int_sel, service; pcap_isr_work() local
185 ezx_pcap_read(pcap, PCAP_REG_MSR, &msr); pcap_isr_work()
194 ezx_pcap_write(pcap, PCAP_REG_MSR, isr | msr); pcap_isr_work()
198 service = isr & ~msr; pcap_isr_work()
204 ezx_pcap_write(pcap, PCAP_REG_MSR, pcap->msr); pcap_isr_work()
476 pcap->msr = PCAP_MASK_ALL_INTERRUPT; ezx_pcap_probe()
/linux-4.1.27/arch/arm/kernel/
H A Diwmmxt.S201 msr cpsr_c, r2
235 1: msr cpsr_c, ip @ restore interrupt mode
253 msr cpsr_c, r2
262 msr cpsr_c, ip @ restore interrupt mode
273 msr cpsr_c, ip @ restore interrupt mode
291 msr cpsr_c, r2
301 msr cpsr_c, ip @ restore interrupt mode
311 msr cpsr_c, ip @ restore interrupt mode
358 msr cpsr_c, ip
364 msr cpsr_c, r2 @ restore interrupts
H A Dentry-header.S157 msr psp, r2
178 msr cpsr_c, \rtemp @ switch to the SYS mode
184 msr cpsr_c, \rtemp @ switch back to the SVC mode
190 msr cpsr_c, \rtemp @ switch to the SYS mode
196 msr cpsr_c, \rtemp @ switch back to the SVC mode
218 msr spsr_cxsf, \rpsr
247 msr cpsr_c, #FIQ_MODE | PSR_I_BIT | PSR_F_BIT
250 msr spsr_cxsf, r9
259 msr spsr_cxsf, r1 @ save in spsr_svc
319 msr cpsr_c, r1
344 msr spsr_cxsf, r1 @ save in spsr_svc
H A Dentry-armv.S332 ARM( msr cpsr_c, #ABT_MODE | PSR_I_BIT | PSR_F_BIT )
334 THUMB( msr cpsr_c, r0 )
337 ARM( msr cpsr_c, #SVC_MODE | PSR_I_BIT | PSR_F_BIT )
339 THUMB( msr cpsr_c, r0 )
346 ARM( msr cpsr_c, #ABT_MODE | PSR_I_BIT | PSR_F_BIT )
348 THUMB( msr cpsr_c, r0 )
350 msr spsr_cxsf, r2 @ Restore spsr_abt
351 ARM( msr cpsr_c, #SVC_MODE | PSR_I_BIT | PSR_F_BIT )
353 THUMB( msr cpsr_c, r0 )
1057 msr spsr_cxsf, r0
/linux-4.1.27/arch/arm64/kernel/
H A Dhead.S511 msr sctlr_el2, x0
516 msr sctlr_el1, x0
523 msr hcr_el2, x0
528 msr cnthctl_el2, x0
529 msr cntvoff_el2, xzr // Clear virtual offset
551 msr vpidr_el2, x0
552 msr vmpidr_el2, x1
558 msr sctlr_el1, x0
562 msr cptr_el2, x0 // Disable copro. traps to EL2
565 msr hstr_el2, xzr // Disable CP15 traps to EL2
575 msr mdcr_el2, x0 // all PMU counters from EL1
579 msr vttbr_el2, xzr
584 msr vbar_el2, x0
589 msr spsr_el2, x0
590 msr elr_el2, lr
684 msr vbar_el1, x5
685 msr ttbr0_el1, x25 // load TTBR0
686 msr ttbr1_el1, x26 // load TTBR1
688 msr sctlr_el1, x0
H A Defi-entry.S101 msr sctlr_el2, x0
108 msr sctlr_el1, x0
H A Dprocess.c218 asm ("msr tpidr_el0, xzr"); tls_thread_flush()
229 asm ("msr tpidrro_el0, xzr"); tls_thread_flush()
319 " msr tpidr_el0, %0\n" tls_thread_switch()
320 " msr tpidrro_el0, %1" tls_thread_switch()
H A Dperf_event.c883 asm volatile("msr pmcr_el0, %0" :: "r" (val)); armv8pmu_pmcr_write()
923 asm volatile("msr pmselr_el0, %0" :: "r" (counter)); armv8pmu_select_counter()
950 asm volatile("msr pmccntr_el0, %0" :: "r" (value)); armv8pmu_write_counter()
952 asm volatile("msr pmxevcntr_el0, %0" :: "r" (value)); armv8pmu_write_counter()
959 asm volatile("msr pmxevtyper_el0, %0" :: "r" (val)); armv8pmu_write_evtype()
974 asm volatile("msr pmcntenset_el0, %0" :: "r" (BIT(counter))); armv8pmu_enable_counter()
989 asm volatile("msr pmcntenclr_el0, %0" :: "r" (BIT(counter))); armv8pmu_disable_counter()
1004 asm volatile("msr pmintenset_el1, %0" :: "r" (BIT(counter))); armv8pmu_enable_intens()
1019 asm volatile("msr pmintenclr_el1, %0" :: "r" (BIT(counter))); armv8pmu_disable_intens()
1022 asm volatile("msr pmovsclr_el0, %0" :: "r" (BIT(counter))); armv8pmu_disable_intens()
1036 asm volatile("msr pmovsclr_el0, %0" :: "r" (value)); armv8pmu_getreset_flags()
H A Dcacheinfo.c59 asm volatile("msr csselr_el1, %x0" : : "r" (csselr)); cache_get_ccsidr()
H A Dhyp-stub.S61 msr vbar_el2, x0 // Set vbar_el2
H A Dsys_compat.c97 asm ("msr tpidrro_el0, %0" : : "r" (regs->regs[0])); compat_arm_syscall()
H A Ddebug-monitors.c46 asm volatile("msr mdscr_el1, %0" :: "r" (mdscr)); mdscr_write()
130 asm volatile("msr oslar_el1, %0" : : "r" (0)); clear_os_lock()
H A Dentry.S124 msr sp_el0, x23
134 "mrs x29, contextidr_el1; msr contextidr_el1, x29; 1:", \
139 "msr contextidr_el1, xzr; 1:", \
144 msr elr_el1, x21 // set up the return data
145 msr spsr_el1, x22
/linux-4.1.27/arch/x86/kvm/
H A Dpmu.c61 static inline struct kvm_pmc *get_gp_pmc(struct kvm_pmu *pmu, u32 msr, get_gp_pmc() argument
64 if (msr >= base && msr < base + pmu->nr_arch_gp_counters) get_gp_pmc()
65 return &pmu->gp_counters[msr - base]; get_gp_pmc()
69 static inline struct kvm_pmc *get_fixed_pmc(struct kvm_pmu *pmu, u32 msr) get_fixed_pmc() argument
72 if (msr >= base && msr < base + pmu->nr_arch_fixed_counters) get_fixed_pmc()
73 return &pmu->fixed_counters[msr - base]; get_fixed_pmc()
323 bool kvm_pmu_msr(struct kvm_vcpu *vcpu, u32 msr) kvm_pmu_msr() argument
328 switch (msr) { kvm_pmu_msr()
336 ret = get_gp_pmc(pmu, msr, MSR_IA32_PERFCTR0) kvm_pmu_msr()
337 || get_gp_pmc(pmu, msr, MSR_P6_EVNTSEL0) kvm_pmu_msr()
338 || get_fixed_pmc(pmu, msr); kvm_pmu_msr()
H A Dlapic.h85 int kvm_x2apic_msr_write(struct kvm_vcpu *vcpu, u32 msr, u64 data);
86 int kvm_x2apic_msr_read(struct kvm_vcpu *vcpu, u32 msr, u64 *data);
88 int kvm_hv_vapic_msr_write(struct kvm_vcpu *vcpu, u32 msr, u64 data);
89 int kvm_hv_vapic_msr_read(struct kvm_vcpu *vcpu, u32 msr, u64 *data);
H A Dx86.c58 #include <asm/msr.h>
202 static void shared_msr_update(unsigned slot, u32 msr) shared_msr_update() argument
214 rdmsrl_safe(msr, &value); shared_msr_update()
219 void kvm_define_shared_msr(unsigned slot, u32 msr) kvm_define_shared_msr() argument
224 shared_msrs_global.msrs[slot] = msr; kvm_define_shared_msr()
920 * List of msr numbers which we expose to userspace through KVM_GET_MSRS
1010 * Writes msr value into into the appropriate "register".
1014 int kvm_set_msr(struct kvm_vcpu *vcpu, struct msr_data *msr) kvm_set_msr() argument
1016 switch (msr->index) { kvm_set_msr()
1022 if (is_noncanonical_address(msr->data)) kvm_set_msr()
1039 msr->data = get_canonical(msr->data); kvm_set_msr()
1041 return kvm_x86_ops->set_msr(vcpu, msr); kvm_set_msr()
1050 struct msr_data msr; do_set_msr() local
1052 msr.data = *data; do_set_msr()
1053 msr.index = index; do_set_msr()
1054 msr.host_initiated = true; do_set_msr()
1055 return kvm_set_msr(vcpu, &msr); do_set_msr()
1293 void kvm_write_tsc(struct kvm_vcpu *vcpu, struct msr_data *msr) kvm_write_tsc() argument
1301 u64 data = msr->data; kvm_write_tsc()
1400 if (guest_cpuid_has_tsc_adjust(vcpu) && !msr->host_initiated) kvm_write_tsc()
1775 static bool msr_mtrr_valid(unsigned msr) msr_mtrr_valid() argument
1777 switch (msr) { msr_mtrr_valid()
1809 bool kvm_mtrr_valid(struct kvm_vcpu *vcpu, u32 msr, u64 data) kvm_mtrr_valid() argument
1814 if (!msr_mtrr_valid(msr)) kvm_mtrr_valid()
1817 if (msr == MSR_IA32_CR_PAT) { kvm_mtrr_valid()
1822 } else if (msr == MSR_MTRRdefType) { kvm_mtrr_valid()
1826 } else if (msr >= MSR_MTRRfix64K_00000 && msr <= MSR_MTRRfix4K_F8000) { kvm_mtrr_valid()
1834 WARN_ON(!(msr >= 0x200 && msr < 0x200 + 2 * KVM_NR_VAR_MTRR)); kvm_mtrr_valid()
1837 if ((msr & 1) == 0) { kvm_mtrr_valid()
1854 static int set_msr_mtrr(struct kvm_vcpu *vcpu, u32 msr, u64 data) set_msr_mtrr() argument
1858 if (!kvm_mtrr_valid(vcpu, msr, data)) set_msr_mtrr()
1861 if (msr == MSR_MTRRdefType) { set_msr_mtrr()
1864 } else if (msr == MSR_MTRRfix64K_00000) set_msr_mtrr()
1866 else if (msr == MSR_MTRRfix16K_80000 || msr == MSR_MTRRfix16K_A0000) set_msr_mtrr()
1867 p[1 + msr - MSR_MTRRfix16K_80000] = data; set_msr_mtrr()
1868 else if (msr >= MSR_MTRRfix4K_C0000 && msr <= MSR_MTRRfix4K_F8000) set_msr_mtrr()
1869 p[3 + msr - MSR_MTRRfix4K_C0000] = data; set_msr_mtrr()
1870 else if (msr == MSR_IA32_CR_PAT) set_msr_mtrr()
1876 idx = (msr - 0x200) / 2; set_msr_mtrr()
1877 is_mtrr_mask = msr - 0x200 - 2 * idx; set_msr_mtrr()
1891 static int set_msr_mce(struct kvm_vcpu *vcpu, u32 msr, u64 data) set_msr_mce() argument
1896 switch (msr) { set_msr_mce()
1908 if (msr >= MSR_IA32_MC0_CTL && set_msr_mce()
1909 msr < MSR_IA32_MCx_CTL(bank_num)) { set_msr_mce()
1910 u32 offset = msr - MSR_IA32_MC0_CTL; set_msr_mce()
1963 static bool kvm_hv_msr_partition_wide(u32 msr) kvm_hv_msr_partition_wide() argument
1966 switch (msr) { kvm_hv_msr_partition_wide()
1978 static int set_msr_hyperv_pw(struct kvm_vcpu *vcpu, u32 msr, u64 data) set_msr_hyperv_pw() argument
1982 switch (msr) { set_msr_hyperv_pw()
2029 "data 0x%llx\n", msr, data); set_msr_hyperv_pw()
2035 static int set_msr_hyperv(struct kvm_vcpu *vcpu, u32 msr, u64 data) set_msr_hyperv() argument
2037 switch (msr) { set_msr_hyperv()
2068 "data 0x%llx\n", msr, data); set_msr_hyperv()
2139 u32 msr = msr_info->index; kvm_set_msr_common() local
2142 switch (msr) { kvm_set_msr_common()
2184 return set_msr_mtrr(vcpu, msr, data); kvm_set_msr_common()
2188 return kvm_x2apic_msr_write(vcpu, msr, data); kvm_set_msr_common()
2217 bool tmp = (msr == MSR_KVM_SYSTEM_TIME); kvm_set_msr_common()
2277 return set_msr_mce(vcpu, msr, data); kvm_set_msr_common()
2292 "0x%x data 0x%llx\n", msr, data); kvm_set_msr_common()
2302 "0x%x data 0x%llx\n", msr, data); kvm_set_msr_common()
2309 if (kvm_pmu_msr(vcpu, msr)) kvm_set_msr_common()
2314 "0x%x data 0x%llx\n", msr, data); kvm_set_msr_common()
2327 if (kvm_hv_msr_partition_wide(msr)) { kvm_set_msr_common()
2330 r = set_msr_hyperv_pw(vcpu, msr, data); kvm_set_msr_common()
2334 return set_msr_hyperv(vcpu, msr, data); kvm_set_msr_common()
2340 vcpu_unimpl(vcpu, "ignored wrmsr: 0x%x data %llx\n", msr, data); kvm_set_msr_common()
2353 if (msr && (msr == vcpu->kvm->arch.xen_hvm_config.msr)) kvm_set_msr_common()
2355 if (kvm_pmu_msr(vcpu, msr)) kvm_set_msr_common()
2359 msr, data); kvm_set_msr_common()
2363 msr, data); kvm_set_msr_common()
2373 * Reads an msr value (of 'msr_index') into 'pdata'.
2383 static int get_msr_mtrr(struct kvm_vcpu *vcpu, u32 msr, u64 *pdata) get_msr_mtrr() argument
2387 if (!msr_mtrr_valid(msr)) get_msr_mtrr()
2390 if (msr == MSR_MTRRdefType) get_msr_mtrr()
2393 else if (msr == MSR_MTRRfix64K_00000) get_msr_mtrr()
2395 else if (msr == MSR_MTRRfix16K_80000 || msr == MSR_MTRRfix16K_A0000) get_msr_mtrr()
2396 *pdata = p[1 + msr - MSR_MTRRfix16K_80000]; get_msr_mtrr()
2397 else if (msr >= MSR_MTRRfix4K_C0000 && msr <= MSR_MTRRfix4K_F8000) get_msr_mtrr()
2398 *pdata = p[3 + msr - MSR_MTRRfix4K_C0000]; get_msr_mtrr()
2399 else if (msr == MSR_IA32_CR_PAT) get_msr_mtrr()
2405 idx = (msr - 0x200) / 2; get_msr_mtrr()
2406 is_mtrr_mask = msr - 0x200 - 2 * idx; get_msr_mtrr()
2419 static int get_msr_mce(struct kvm_vcpu *vcpu, u32 msr, u64 *pdata) get_msr_mce() argument
2425 switch (msr) { get_msr_mce()
2442 if (msr >= MSR_IA32_MC0_CTL && get_msr_mce()
2443 msr < MSR_IA32_MCx_CTL(bank_num)) { get_msr_mce()
2444 u32 offset = msr - MSR_IA32_MC0_CTL; get_msr_mce()
2454 static int get_msr_hyperv_pw(struct kvm_vcpu *vcpu, u32 msr, u64 *pdata) get_msr_hyperv_pw() argument
2459 switch (msr) { get_msr_hyperv_pw()
2475 vcpu_unimpl(vcpu, "Hyper-V unhandled rdmsr: 0x%x\n", msr); get_msr_hyperv_pw()
2483 static int get_msr_hyperv(struct kvm_vcpu *vcpu, u32 msr, u64 *pdata) get_msr_hyperv() argument
2487 switch (msr) { get_msr_hyperv()
2509 vcpu_unimpl(vcpu, "Hyper-V unhandled rdmsr: 0x%x\n", msr); get_msr_hyperv()
2516 int kvm_get_msr_common(struct kvm_vcpu *vcpu, u32 msr, u64 *pdata) kvm_get_msr_common() argument
2520 switch (msr) { kvm_get_msr_common()
2549 if (kvm_pmu_msr(vcpu, msr)) kvm_get_msr_common()
2550 return kvm_pmu_get_msr(vcpu, msr, pdata); kvm_get_msr_common()
2560 return get_msr_mtrr(vcpu, msr, pdata); kvm_get_msr_common()
2582 return kvm_x2apic_msr_read(vcpu, msr, pdata); kvm_get_msr_common()
2625 return get_msr_mce(vcpu, msr, pdata); kvm_get_msr_common()
2639 if (kvm_hv_msr_partition_wide(msr)) { kvm_get_msr_common()
2642 r = get_msr_hyperv_pw(vcpu, msr, pdata); kvm_get_msr_common()
2646 return get_msr_hyperv(vcpu, msr, pdata); kvm_get_msr_common()
2672 if (kvm_pmu_msr(vcpu, msr)) kvm_get_msr_common()
2673 return kvm_pmu_get_msr(vcpu, msr, pdata); kvm_get_msr_common()
2675 vcpu_unimpl(vcpu, "unhandled rdmsr: 0x%x\n", msr); kvm_get_msr_common()
2678 vcpu_unimpl(vcpu, "ignored rdmsr: 0x%x\n", msr); kvm_get_msr_common()
4957 struct msr_data msr; emulator_set_msr() local
4959 msr.data = data; emulator_set_msr()
4960 msr.index = msr_index; emulator_set_msr()
4961 msr.host_initiated = false; emulator_set_msr()
4962 return kvm_set_msr(emul_to_vcpu(ctxt), &msr); emulator_set_msr()
7116 struct msr_data msr; kvm_arch_vcpu_postcreate() local
7121 msr.data = 0x0; kvm_arch_vcpu_postcreate()
7122 msr.index = MSR_IA32_TSC; kvm_arch_vcpu_postcreate()
7123 msr.host_initiated = true; kvm_arch_vcpu_postcreate()
7124 kvm_write_tsc(vcpu, &msr); kvm_arch_vcpu_postcreate()
H A Dvmx.c1188 static int __find_msr_index(struct vcpu_vmx *vmx, u32 msr) __find_msr_index() argument
1193 if (vmx_msr_index[vmx->guest_msrs[i].index] == msr) __find_msr_index()
1224 static struct shared_msr_entry *find_msr_entry(struct vcpu_vmx *vmx, u32 msr) find_msr_entry() argument
1228 i = __find_msr_index(vmx, msr); find_msr_entry()
1600 static void clear_atomic_switch_msr(struct vcpu_vmx *vmx, unsigned msr) clear_atomic_switch_msr() argument
1605 switch (msr) { clear_atomic_switch_msr()
1625 if (m->guest[i].index == msr) clear_atomic_switch_msr()
1648 static void add_atomic_switch_msr(struct vcpu_vmx *vmx, unsigned msr, add_atomic_switch_msr() argument
1654 switch (msr) { add_atomic_switch_msr()
1687 if (m->guest[i].index == msr) add_atomic_switch_msr()
1691 printk_once(KERN_WARNING "Not enough msr switch entries. " add_atomic_switch_msr()
1692 "Can't add msr %x\n", msr); add_atomic_switch_msr()
1700 m->guest[i].index = msr; add_atomic_switch_msr()
1702 m->host[i].index = msr; add_atomic_switch_msr()
2639 * Reads an msr value (of 'msr_index') into 'pdata'.
2646 struct shared_msr_entry *msr; vmx_get_msr() local
2704 msr = find_msr_entry(to_vmx(vcpu), msr_index); vmx_get_msr()
2705 if (msr) { vmx_get_msr()
2706 data = msr->data; vmx_get_msr()
2719 * Writes msr value into into the appropriate "register".
2726 struct shared_msr_entry *msr; vmx_set_msr() local
2814 msr = find_msr_entry(vmx, msr_index); vmx_set_msr()
2815 if (msr) { vmx_set_msr()
2816 u64 old_msr_data = msr->data; vmx_set_msr()
2817 msr->data = data; vmx_set_msr()
2818 if (msr - vmx->guest_msrs < vmx->save_nmsrs) { vmx_set_msr()
2820 ret = kvm_set_shared_msr(msr->index, msr->data, vmx_set_msr()
2821 msr->mask); vmx_set_msr()
2824 msr->data = old_msr_data; vmx_set_msr()
2860 u64 msr; vmx_disabled_by_bios() local
2862 rdmsrl(MSR_IA32_FEATURE_CONTROL, msr); vmx_disabled_by_bios()
2863 if (msr & FEATURE_CONTROL_LOCKED) { vmx_disabled_by_bios()
2865 if (!(msr & FEATURE_CONTROL_VMXON_ENABLED_INSIDE_SMX) vmx_disabled_by_bios()
2869 if (!(msr & FEATURE_CONTROL_VMXON_ENABLED_OUTSIDE_SMX) vmx_disabled_by_bios()
2870 && (msr & FEATURE_CONTROL_VMXON_ENABLED_INSIDE_SMX) vmx_disabled_by_bios()
2877 if (!(msr & FEATURE_CONTROL_VMXON_ENABLED_OUTSIDE_SMX) vmx_disabled_by_bios()
2966 u32 msr, u32 *result) adjust_vmx_controls()
2971 rdmsr(msr, vmx_msr_low, vmx_msr_high); adjust_vmx_controls()
2984 static __init bool allow_1_setting(u32 msr, u32 ctl) allow_1_setting() argument
2988 rdmsr(msr, vmx_msr_low, vmx_msr_high); allow_1_setting()
3139 * msr load mechanism to switch IA32_PERF_GLOBAL_CTRL. setup_vmcs_config()
3420 struct shared_msr_entry *msr = find_msr_entry(vmx, MSR_EFER); vmx_set_efer() local
3422 if (!msr) vmx_set_efer()
3427 * of this msr depends on is_long_mode(). vmx_set_efer()
3433 msr->data = efer; vmx_set_efer()
3437 msr->data = efer & ~EFER_LME; vmx_set_efer()
4213 u32 msr, int type) __vmx_disable_intercept_for_msr()
4225 if (msr <= 0x1fff) { __vmx_disable_intercept_for_msr()
4228 __clear_bit(msr, msr_bitmap + 0x000 / f); __vmx_disable_intercept_for_msr()
4232 __clear_bit(msr, msr_bitmap + 0x800 / f); __vmx_disable_intercept_for_msr()
4234 } else if ((msr >= 0xc0000000) && (msr <= 0xc0001fff)) { __vmx_disable_intercept_for_msr()
4235 msr &= 0x1fff; __vmx_disable_intercept_for_msr()
4238 __clear_bit(msr, msr_bitmap + 0x400 / f); __vmx_disable_intercept_for_msr()
4242 __clear_bit(msr, msr_bitmap + 0xc00 / f); __vmx_disable_intercept_for_msr()
4248 u32 msr, int type) __vmx_enable_intercept_for_msr()
4260 if (msr <= 0x1fff) { __vmx_enable_intercept_for_msr()
4263 __set_bit(msr, msr_bitmap + 0x000 / f); __vmx_enable_intercept_for_msr()
4267 __set_bit(msr, msr_bitmap + 0x800 / f); __vmx_enable_intercept_for_msr()
4269 } else if ((msr >= 0xc0000000) && (msr <= 0xc0001fff)) { __vmx_enable_intercept_for_msr()
4270 msr &= 0x1fff; __vmx_enable_intercept_for_msr()
4273 __set_bit(msr, msr_bitmap + 0x400 / f); __vmx_enable_intercept_for_msr()
4277 __set_bit(msr, msr_bitmap + 0xc00 / f); __vmx_enable_intercept_for_msr()
4283 * If a msr is allowed by L0, we should check whether it is allowed by L1.
4288 u32 msr, int type) nested_vmx_disable_intercept_for_msr()
4302 if (msr <= 0x1fff) { nested_vmx_disable_intercept_for_msr()
4304 !test_bit(msr, msr_bitmap_l1 + 0x000 / f)) nested_vmx_disable_intercept_for_msr()
4306 __clear_bit(msr, msr_bitmap_nested + 0x000 / f); nested_vmx_disable_intercept_for_msr()
4309 !test_bit(msr, msr_bitmap_l1 + 0x800 / f)) nested_vmx_disable_intercept_for_msr()
4311 __clear_bit(msr, msr_bitmap_nested + 0x800 / f); nested_vmx_disable_intercept_for_msr()
4313 } else if ((msr >= 0xc0000000) && (msr <= 0xc0001fff)) { nested_vmx_disable_intercept_for_msr()
4314 msr &= 0x1fff; nested_vmx_disable_intercept_for_msr()
4316 !test_bit(msr, msr_bitmap_l1 + 0x400 / f)) nested_vmx_disable_intercept_for_msr()
4318 __clear_bit(msr, msr_bitmap_nested + 0x400 / f); nested_vmx_disable_intercept_for_msr()
4321 !test_bit(msr, msr_bitmap_l1 + 0xc00 / f)) nested_vmx_disable_intercept_for_msr()
4323 __clear_bit(msr, msr_bitmap_nested + 0xc00 / f); nested_vmx_disable_intercept_for_msr()
4328 static void vmx_disable_intercept_for_msr(u32 msr, bool longmode_only) vmx_disable_intercept_for_msr() argument
4332 msr, MSR_TYPE_R | MSR_TYPE_W); vmx_disable_intercept_for_msr()
4334 msr, MSR_TYPE_R | MSR_TYPE_W); vmx_disable_intercept_for_msr()
4337 static void vmx_enable_intercept_msr_read_x2apic(u32 msr) vmx_enable_intercept_msr_read_x2apic() argument
4340 msr, MSR_TYPE_R); vmx_enable_intercept_msr_read_x2apic()
4342 msr, MSR_TYPE_R); vmx_enable_intercept_msr_read_x2apic()
4345 static void vmx_disable_intercept_msr_read_x2apic(u32 msr) vmx_disable_intercept_msr_read_x2apic() argument
4348 msr, MSR_TYPE_R); vmx_disable_intercept_msr_read_x2apic()
4350 msr, MSR_TYPE_R); vmx_disable_intercept_msr_read_x2apic()
4353 static void vmx_disable_intercept_msr_write_x2apic(u32 msr) vmx_disable_intercept_msr_write_x2apic() argument
4356 msr, MSR_TYPE_W); vmx_disable_intercept_msr_write_x2apic()
4358 msr, MSR_TYPE_W); vmx_disable_intercept_msr_write_x2apic()
5518 struct msr_data msr; handle_wrmsr() local
5523 msr.data = data; handle_wrmsr()
5524 msr.index = ecx; handle_wrmsr()
5525 msr.host_initiated = false; handle_wrmsr()
5526 if (kvm_set_msr(vcpu, &msr) != 0) { handle_wrmsr()
6030 int r = -ENOMEM, i, msr; hardware_setup() local
6171 for (msr = 0x800; msr <= 0x8ff; msr++) hardware_setup()
6172 vmx_disable_intercept_msr_read_x2apic(msr); hardware_setup()
8160 clear_atomic_switch_msr(vmx, msrs[i].msr); atomic_switch_perf_msrs()
8162 add_atomic_switch_msr(vmx, msrs[i].msr, msrs[i].guest, atomic_switch_perf_msrs()
8786 int msr; nested_vmx_merge_msr_bitmap() local
8807 for (msr = 0x800; msr <= 0x8ff; msr++) nested_vmx_merge_msr_bitmap()
8811 msr, MSR_TYPE_R); nested_vmx_merge_msr_bitmap()
8837 for (msr = 0x800; msr <= 0x8ff; msr++) nested_vmx_merge_msr_bitmap()
8840 msr, nested_vmx_merge_msr_bitmap()
8982 * Load guest's/host's msr at nested entry/exit.
8989 struct msr_data msr; nested_vmx_load_msr() local
8991 msr.host_initiated = false; nested_vmx_load_msr()
9006 msr.index = e.index; nested_vmx_load_msr()
9007 msr.data = e.value; nested_vmx_load_msr()
9008 if (kvm_set_msr(vcpu, &msr)) { nested_vmx_load_msr()
2965 adjust_vmx_controls(u32 ctl_min, u32 ctl_opt, u32 msr, u32 *result) adjust_vmx_controls() argument
4212 __vmx_disable_intercept_for_msr(unsigned long *msr_bitmap, u32 msr, int type) __vmx_disable_intercept_for_msr() argument
4247 __vmx_enable_intercept_for_msr(unsigned long *msr_bitmap, u32 msr, int type) __vmx_enable_intercept_for_msr() argument
4286 nested_vmx_disable_intercept_for_msr(unsigned long *msr_bitmap_l1, unsigned long *msr_bitmap_nested, u32 msr, int type) nested_vmx_disable_intercept_for_msr() argument
H A Dx86.h153 void kvm_write_tsc(struct kvm_vcpu *vcpu, struct msr_data *msr);
163 bool kvm_mtrr_valid(struct kvm_vcpu *vcpu, u32 msr, u64 data);
H A Dsvm.c426 static u32 svm_msrpm_offset(u32 msr) svm_msrpm_offset() argument
432 if (msr < msrpm_ranges[i] || svm_msrpm_offset()
433 msr >= msrpm_ranges[i] + MSRS_IN_RANGE) svm_msrpm_offset()
436 offset = (msr - msrpm_ranges[i]) / 4; /* 4 msrs per u8 */ svm_msrpm_offset()
762 static void set_msr_interception(u32 *msrpm, unsigned msr, set_msr_interception() argument
773 WARN_ON(!valid_msr_intercept(msr)); set_msr_interception()
775 offset = svm_msrpm_offset(msr); set_msr_interception()
776 bit_read = 2 * (msr & 0x0f); set_msr_interception()
777 bit_write = 2 * (msr & 0x0f) + 1; set_msr_interception()
2164 u32 offset, msr, value; nested_svm_exit_handled_msr() local
2170 msr = svm->vcpu.arch.regs[VCPU_REGS_RCX]; nested_svm_exit_handled_msr()
2171 offset = svm_msrpm_offset(msr); nested_svm_exit_handled_msr()
2173 mask = 1 << ((2 * (msr & 0xf)) + write); nested_svm_exit_handled_msr()
2433 * This function merges the msr permission bitmaps of kvm and the nested_svm_vmrun_msrpm()
2435 * the kvm msr permission bitmap may contain zero bits nested_svm_vmrun_msrpm()
3188 static int svm_set_msr(struct kvm_vcpu *vcpu, struct msr_data *msr) svm_set_msr() argument
3192 u32 ecx = msr->index; svm_set_msr()
3193 u64 data = msr->data; svm_set_msr()
3196 kvm_write_tsc(vcpu, msr); svm_set_msr()
3251 return kvm_set_msr_common(vcpu, msr); svm_set_msr()
3258 struct msr_data msr; wrmsr_interception() local
3262 msr.data = data; wrmsr_interception()
3263 msr.index = ecx; wrmsr_interception()
3264 msr.host_initiated = false; wrmsr_interception()
3267 if (kvm_set_msr(&svm->vcpu, &msr)) { wrmsr_interception()
/linux-4.1.27/drivers/tty/serial/
H A Dvr41xx_siu.c223 uint8_t msr; siu_get_mctrl() local
226 msr = siu_read(port, UART_MSR); siu_get_mctrl()
227 if (msr & UART_MSR_DCD) siu_get_mctrl()
229 if (msr & UART_MSR_RI) siu_get_mctrl()
231 if (msr & UART_MSR_DSR) siu_get_mctrl()
233 if (msr & UART_MSR_CTS) siu_get_mctrl()
373 uint8_t msr; check_modem_status() local
375 msr = siu_read(port, UART_MSR); check_modem_status()
376 if ((msr & UART_MSR_ANY_DELTA) == 0) check_modem_status()
378 if (msr & UART_MSR_DDCD) check_modem_status()
379 uart_handle_dcd_change(port, msr & UART_MSR_DCD); check_modem_status()
380 if (msr & UART_MSR_TERI) check_modem_status()
382 if (msr & UART_MSR_DDSR) check_modem_status()
384 if (msr & UART_MSR_DCTS) check_modem_status()
385 uart_handle_cts_change(port, msr & UART_MSR_CTS); check_modem_status()
741 uint8_t lsr, msr; wait_for_xmitr() local
756 msr = siu_read(port, UART_MSR); wait_for_xmitr()
757 if ((msr & UART_MSR_CTS) != 0) wait_for_xmitr()
H A Dmen_z135_uart.c187 u8 msr; men_z135_handle_modem_status() local
189 msr = (uart->stat_reg >> 8) & 0xff; men_z135_handle_modem_status()
191 if (msr & MEN_Z135_MSR_DDCD) men_z135_handle_modem_status()
193 msr & MEN_Z135_MSR_DCD); men_z135_handle_modem_status()
194 if (msr & MEN_Z135_MSR_DCTS) men_z135_handle_modem_status()
196 msr & MEN_Z135_MSR_CTS); men_z135_handle_modem_status()
528 u8 msr; men_z135_get_mctrl() local
530 msr = ioread8(port->membase + MEN_Z135_STAT_REG + 1); men_z135_get_mctrl()
532 if (msr & MEN_Z135_MSR_CTS) men_z135_get_mctrl()
534 if (msr & MEN_Z135_MSR_DSR) men_z135_get_mctrl()
536 if (msr & MEN_Z135_MSR_RI) men_z135_get_mctrl()
538 if (msr & MEN_Z135_MSR_DCD) men_z135_get_mctrl()
H A Dserial-tegra.c650 unsigned long msr; tegra_uart_handle_modem_signal_change() local
652 msr = tegra_uart_read(tup, UART_MSR); tegra_uart_handle_modem_signal_change()
653 if (!(msr & UART_MSR_ANY_DELTA)) tegra_uart_handle_modem_signal_change()
656 if (msr & UART_MSR_TERI) tegra_uart_handle_modem_signal_change()
658 if (msr & UART_MSR_DDSR) tegra_uart_handle_modem_signal_change()
661 if (msr & UART_MSR_DDCD) tegra_uart_handle_modem_signal_change()
662 uart_handle_dcd_change(&tup->uport, msr & UART_MSR_DCD); tegra_uart_handle_modem_signal_change()
664 if (msr & UART_MSR_DCTS) tegra_uart_handle_modem_signal_change()
665 uart_handle_cts_change(&tup->uport, msr & UART_MSR_CTS); tegra_uart_handle_modem_signal_change()
784 unsigned long msr; tegra_uart_hw_deinit() local
792 msr = tegra_uart_read(tup, UART_MSR); tegra_uart_hw_deinit()
794 if ((mcr & TEGRA_UART_MCR_CTS_EN) && (msr & UART_MSR_CTS)) tegra_uart_hw_deinit()
804 msr = tegra_uart_read(tup, UART_MSR); tegra_uart_hw_deinit()
807 (msr & UART_MSR_CTS)) tegra_uart_hw_deinit()
H A Dpnx8xxx_uart.c331 unsigned int msr; pnx8xxx_get_mctrl() local
335 msr = serial_in(sport, PNX8XXX_MCR); pnx8xxx_get_mctrl()
337 mctrl |= msr & PNX8XXX_UART_MCR_CTS ? TIOCM_CTS : 0; pnx8xxx_get_mctrl()
338 mctrl |= msr & PNX8XXX_UART_MCR_DCD ? TIOCM_CAR : 0; pnx8xxx_get_mctrl()
347 unsigned int msr; pnx8xxx_set_mctrl()
/linux-4.1.27/arch/powerpc/sysdev/
H A Duic.c202 u32 msr; uic_irq_cascade() local
213 msr = mfdcr(uic->dcrbase + UIC_MSR); uic_irq_cascade()
214 if (!msr) /* spurious interrupt */ uic_irq_cascade()
217 src = 32 - ffs(msr); uic_irq_cascade()
325 u32 msr; uic_get_irq() local
330 msr = mfdcr(primary_uic->dcrbase + UIC_MSR); uic_get_irq()
331 src = 32 - ffs(msr); uic_get_irq()
/linux-4.1.27/drivers/usb/serial/
H A Dmct_u232.c310 unsigned char *msr) mct_u232_get_modem_stat()
317 *msr = 0; mct_u232_get_modem_stat()
327 *msr = 0; mct_u232_get_modem_stat()
329 *msr = buf[0]; mct_u232_get_modem_stat()
331 dev_dbg(&port->dev, "get_modem_stat: 0x%x\n", *msr); mct_u232_get_modem_stat()
337 unsigned char msr) mct_u232_msr_to_icount()
340 if (msr & MCT_U232_MSR_DDSR) mct_u232_msr_to_icount()
342 if (msr & MCT_U232_MSR_DCTS) mct_u232_msr_to_icount()
344 if (msr & MCT_U232_MSR_DRI) mct_u232_msr_to_icount()
346 if (msr & MCT_U232_MSR_DCD) mct_u232_msr_to_icount()
351 unsigned int *control_state, unsigned char msr) mct_u232_msr_to_state()
354 if (msr & MCT_U232_MSR_DSR) mct_u232_msr_to_state()
358 if (msr & MCT_U232_MSR_CTS) mct_u232_msr_to_state()
362 if (msr & MCT_U232_MSR_RI) mct_u232_msr_to_state()
366 if (msr & MCT_U232_MSR_CD) mct_u232_msr_to_state()
370 dev_dbg(&port->dev, "msr_to_state: msr=0x%x ==> state=0x%x\n", msr, *control_state); mct_u232_msr_to_state()
309 mct_u232_get_modem_stat(struct usb_serial_port *port, unsigned char *msr) mct_u232_get_modem_stat() argument
336 mct_u232_msr_to_icount(struct async_icount *icount, unsigned char msr) mct_u232_msr_to_icount() argument
350 mct_u232_msr_to_state(struct usb_serial_port *port, unsigned int *control_state, unsigned char msr) mct_u232_msr_to_state() argument
H A Df81232.c233 case 0x00: /* msr change */ f81232_update_line_status()
505 u8 mcr, msr; f81232_tiocmget() local
512 msr = port_priv->modem_status; f81232_tiocmget()
517 (msr & UART_MSR_CTS ? TIOCM_CTS : 0) | f81232_tiocmget()
518 (msr & UART_MSR_DCD ? TIOCM_CAR : 0) | f81232_tiocmget()
519 (msr & UART_MSR_RI ? TIOCM_RI : 0) | f81232_tiocmget()
520 (msr & UART_MSR_DSR ? TIOCM_DSR : 0); f81232_tiocmget()
578 u8 msr; f81232_carrier_raised() local
582 msr = priv->modem_status; f81232_carrier_raised()
585 if (msr & UART_MSR_DCD) f81232_carrier_raised()
H A Dti_usb_3410_5052.c123 static void ti_handle_new_msr(struct ti_port *tport, __u8 msr);
842 unsigned int msr; ti_tiocmget() local
850 msr = tport->tp_msr; ti_tiocmget()
857 | ((msr & TI_MSR_CTS) ? TIOCM_CTS : 0) ti_tiocmget()
858 | ((msr & TI_MSR_CD) ? TIOCM_CAR : 0) ti_tiocmget()
859 | ((msr & TI_MSR_RI) ? TIOCM_RI : 0) ti_tiocmget()
860 | ((msr & TI_MSR_DSR) ? TIOCM_DSR : 0); ti_tiocmget()
934 __u8 msr; ti_interrupt_callback() local
986 msr = data[1]; ti_interrupt_callback()
987 dev_dbg(dev, "%s - port %d, msr 0x%02X\n", __func__, port_number, msr); ti_interrupt_callback()
988 ti_handle_new_msr(tport, msr); ti_interrupt_callback()
1267 static void ti_handle_new_msr(struct ti_port *tport, __u8 msr) ti_handle_new_msr() argument
1273 dev_dbg(&tport->tp_port->dev, "%s - msr 0x%02X\n", __func__, msr); ti_handle_new_msr()
1275 if (msr & TI_MSR_DELTA_MASK) { ti_handle_new_msr()
1278 if (msr & TI_MSR_DELTA_CTS) ti_handle_new_msr()
1280 if (msr & TI_MSR_DELTA_DSR) ti_handle_new_msr()
1282 if (msr & TI_MSR_DELTA_CD) ti_handle_new_msr()
1284 if (msr & TI_MSR_DELTA_RI) ti_handle_new_msr()
1290 tport->tp_msr = msr & TI_MSR_MASK; ti_handle_new_msr()
1295 if (msr & TI_MSR_CTS) ti_handle_new_msr()
H A Dssu100.c441 static void ssu100_update_msr(struct usb_serial_port *port, u8 msr) ssu100_update_msr() argument
447 priv->shadowMSR = msr; ssu100_update_msr()
450 if (msr & UART_MSR_ANY_DELTA) { ssu100_update_msr()
452 if (msr & UART_MSR_DCTS) ssu100_update_msr()
454 if (msr & UART_MSR_DDSR) ssu100_update_msr()
456 if (msr & UART_MSR_DDCD) ssu100_update_msr()
458 if (msr & UART_MSR_TERI) ssu100_update_msr()
H A Dark3116.c77 __u32 msr; /* modem status register value */ member in struct:ark3116_private
376 priv->msr = ark3116_read_reg(serial, UART_MSR, buf); ark3116_open()
446 status = priv->msr; ark3116_tiocmget()
513 static void ark3116_update_msr(struct usb_serial_port *port, __u8 msr) ark3116_update_msr() argument
519 priv->msr = msr; ark3116_update_msr()
522 if (msr & UART_MSR_ANY_DELTA) { ark3116_update_msr()
524 if (msr & UART_MSR_DCTS) ark3116_update_msr()
526 if (msr & UART_MSR_DDSR) ark3116_update_msr()
528 if (msr & UART_MSR_DDCD) ark3116_update_msr()
530 if (msr & UART_MSR_TERI) ark3116_update_msr()
583 dev_dbg(&port->dev, "%s: msr=%02x\n", ark3116_read_int_callback()
H A Dio_ti.c1430 static void handle_new_msr(struct edgeport_port *edge_port, __u8 msr) handle_new_msr() argument
1435 dev_dbg(&edge_port->port->dev, "%s - %02x\n", __func__, msr); handle_new_msr()
1437 if (msr & (EDGEPORT_MSR_DELTA_CTS | EDGEPORT_MSR_DELTA_DSR | handle_new_msr()
1442 if (msr & EDGEPORT_MSR_DELTA_CTS) handle_new_msr()
1444 if (msr & EDGEPORT_MSR_DELTA_DSR) handle_new_msr()
1446 if (msr & EDGEPORT_MSR_DELTA_CD) handle_new_msr()
1448 if (msr & EDGEPORT_MSR_DELTA_RI) handle_new_msr()
1454 edge_port->shadow_msr = msr & 0xf0; handle_new_msr()
1459 if (msr & EDGEPORT_MSR_CTS) handle_new_msr()
1512 __u8 msr; edge_interrupt_callback() local
1574 msr = data[1]; edge_interrupt_callback()
1576 __func__, port_number, msr); edge_interrupt_callback()
1577 handle_new_msr(edge_port, msr); edge_interrupt_callback()
2294 unsigned int msr; edge_tiocmget() local
2300 msr = edge_port->shadow_msr; edge_tiocmget()
2304 | ((msr & EDGEPORT_MSR_CTS) ? TIOCM_CTS: 0) /* 0x020 */ edge_tiocmget()
2305 | ((msr & EDGEPORT_MSR_CD) ? TIOCM_CAR: 0) /* 0x040 */ edge_tiocmget()
2306 | ((msr & EDGEPORT_MSR_RI) ? TIOCM_RI: 0) /* 0x080 */ edge_tiocmget()
2307 | ((msr & EDGEPORT_MSR_DSR) ? TIOCM_DSR: 0); /* 0x100 */ edge_tiocmget()
H A Dspcp8x5.c252 u8 msr; spcp8x5_carrier_raised() local
255 ret = spcp8x5_get_msr(port, &msr); spcp8x5_carrier_raised()
256 if (ret || msr & MSR_STATUS_LINE_DCD) spcp8x5_carrier_raised()
/linux-4.1.27/arch/x86/kernel/cpu/mcheck/
H A Dwinchip.c13 #include <asm/msr.h>
H A Dp5.c14 #include <asm/msr.h>
H A Dmce.c49 #include <asm/msr.h>
373 static int msr_to_offset(u32 msr) msr_to_offset() argument
377 if (msr == mca_cfg.rip_msr) msr_to_offset()
379 if (msr == MSR_IA32_MCx_STATUS(bank)) msr_to_offset()
381 if (msr == MSR_IA32_MCx_ADDR(bank)) msr_to_offset()
383 if (msr == MSR_IA32_MCx_MISC(bank)) msr_to_offset()
385 if (msr == MSR_IA32_MCG_STATUS) msr_to_offset()
391 static u64 mce_rdmsrl(u32 msr) mce_rdmsrl() argument
396 int offset = msr_to_offset(msr); mce_rdmsrl()
403 if (rdmsrl_safe(msr, &v)) { mce_rdmsrl()
404 WARN_ONCE(1, "mce: Unable to read msr %d!\n", msr); mce_rdmsrl()
416 static void mce_wrmsrl(u32 msr, u64 v) mce_wrmsrl() argument
419 int offset = msr_to_offset(msr); mce_wrmsrl()
425 wrmsrl(msr, v); mce_wrmsrl()
1249 * MSR_IA32_THERMAL_STATUS (Intel) msr.
H A Dmce_amd.c34 #include <asm/msr.h>
119 int msr = (hi & MASK_LVTOFF_HI) >> 20; lvt_off_valid() local
128 if (apic != msr) { lvt_off_valid()
/linux-4.1.27/arch/mips/pci/
H A Dops-loongson2.c185 void _rdmsr(u32 msr, u32 *hi, u32 *lo) _rdmsr() argument
194 loongson_pcibios_write(&bus, devfn, PCI_MSR_ADDR, 4, msr); _rdmsr()
201 void _wrmsr(u32 msr, u32 hi, u32 lo) _wrmsr() argument
210 loongson_pcibios_write(&bus, devfn, PCI_MSR_ADDR, 4, msr); _wrmsr()
/linux-4.1.27/arch/arm/lib/
H A Decard.S18 msr spsr_cxsf, rt
/linux-4.1.27/drivers/staging/dgnc/
H A Ddgnc_cls.h35 * msr : WR MSG - Modem Status Reg
45 u8 msr; member in struct:cls_uart_struct
H A Ddgnc_neo.h37 u8 msr; /* WR MSR - Modem Status Reg */ member in struct:neo_uart_struct
/linux-4.1.27/arch/arm/mach-ep93xx/
H A Dcrunch-bits.S214 msr cpsr_c, r2
246 1: msr cpsr_c, ip @ restore interrupt mode
260 msr cpsr_c, r2
269 msr cpsr_c, ip @ restore interrupt mode
279 msr cpsr_c, ip @ restore interrupt mode
293 msr cpsr_c, r2
302 msr cpsr_c, ip @ restore interrupt mode
312 msr cpsr_c, ip @ restore interrupt mode
/linux-4.1.27/tools/power/cpupower/utils/helpers/
H A Dmsr.c31 sprintf(msr_file_name, "/dev/cpu/%d/msr", cpu); read_msr()
60 sprintf(msr_file_name, "/dev/cpu/%d/msr", cpu); write_msr()
H A Dhelpers.h125 /* Read/Write msr ****************************/
133 /* Read/Write msr ****************************/
181 /* Read/Write msr ****************************/
/linux-4.1.27/drivers/i2c/busses/
H A Di2c-rcar.c268 static int rcar_i2c_irq_send(struct rcar_i2c_priv *priv, u32 msr) rcar_i2c_irq_send() argument
277 if (!(msr & MDE)) rcar_i2c_irq_send()
284 if (msr & MAT) rcar_i2c_irq_send()
329 static int rcar_i2c_irq_recv(struct rcar_i2c_priv *priv, u32 msr) rcar_i2c_irq_recv() argument
338 if (!(msr & MDR)) rcar_i2c_irq_recv()
341 if (msr & MAT) { rcar_i2c_irq_recv()
429 u32 msr; rcar_i2c_irq() local
437 msr = rcar_i2c_read(priv, ICMSR); rcar_i2c_irq()
440 msr &= rcar_i2c_read(priv, ICMIER); rcar_i2c_irq()
441 if (!msr) { rcar_i2c_irq()
447 if (msr & MAL) { rcar_i2c_irq()
453 if (msr & MNR) { rcar_i2c_irq()
462 if (msr & MST) { rcar_i2c_irq()
468 rcar_i2c_flags_set(priv, rcar_i2c_irq_recv(priv, msr)); rcar_i2c_irq()
470 rcar_i2c_flags_set(priv, rcar_i2c_irq_send(priv, msr)); rcar_i2c_irq()
H A Di2c-sh7760.c117 unsigned long msr, fsr, fier, len; sh7760_i2c_irq() local
119 msr = IN32(id, I2CMSR); sh7760_i2c_irq()
123 if (msr & MSR_MAL) { sh7760_i2c_irq()
131 if (msr & MSR_MNR) { sh7760_i2c_irq()
146 msr &= ~MSR_MAT; sh7760_i2c_irq()
152 if (msr & MSR_MST) { sh7760_i2c_irq()
158 if (msr & MSR_MAT) sh7760_i2c_irq()
231 OUT32(id, I2CMSR, ~msr); sh7760_i2c_irq()
/linux-4.1.27/drivers/powercap/
H A Dintel_rapl.c702 /* name, mask, shift, msr index, unit divisor */
759 u32 msr; rapl_read_data_raw() local
766 msr = rd->msrs[rp->id]; rapl_read_data_raw()
767 if (!msr) rapl_read_data_raw()
785 if (rdmsrl_safe_on_cpu(cpu, msr, &value)) { rapl_read_data_raw()
786 pr_debug("failed to read msr 0x%x on cpu %d\n", msr, cpu); rapl_read_data_raw()
806 u32 msr; rapl_write_data_raw() local
813 msr = rd->msrs[rp->id]; rapl_write_data_raw()
814 if (rdmsrl_safe_on_cpu(cpu, msr, &msr_val)) { rapl_write_data_raw()
816 "failed to read msr 0x%x on cpu %d\n", msr, cpu); rapl_write_data_raw()
822 if (wrmsrl_safe_on_cpu(cpu, msr, msr_val)) { rapl_write_data_raw()
824 "failed to write msr 0x%x on cpu %d\n", msr, cpu); rapl_write_data_raw()
1242 unsigned msr; rapl_check_domain() local
1247 msr = MSR_PKG_ENERGY_STATUS; rapl_check_domain()
1250 msr = MSR_PP0_ENERGY_STATUS; rapl_check_domain()
1253 msr = MSR_PP1_ENERGY_STATUS; rapl_check_domain()
1256 msr = MSR_DRAM_ENERGY_STATUS; rapl_check_domain()
1265 if (rdmsrl_safe_on_cpu(cpu, msr, &val) || !val) rapl_check_domain()
/linux-4.1.27/ipc/
H A Dmsg.c188 struct msg_receiver *msr, *t; expunge_all() local
190 list_for_each_entry_safe(msr, t, &msq->q_receivers, r_list) { expunge_all()
191 msr->r_msg = NULL; /* initialize expunge ordering */ expunge_all()
192 wake_up_process(msr->r_tsk); expunge_all()
200 msr->r_msg = ERR_PTR(res); expunge_all()
571 struct msg_receiver *msr, *t; pipelined_send() local
573 list_for_each_entry_safe(msr, t, &msq->q_receivers, r_list) { pipelined_send()
574 if (testmsg(msg, msr->r_msgtype, msr->r_mode) && pipelined_send()
575 !security_msg_queue_msgrcv(msq, msg, msr->r_tsk, pipelined_send()
576 msr->r_msgtype, msr->r_mode)) { pipelined_send()
578 list_del(&msr->r_list); pipelined_send()
579 if (msr->r_maxsize < msg->m_ts) { pipelined_send()
581 msr->r_msg = NULL; pipelined_send()
582 wake_up_process(msr->r_tsk); pipelined_send()
584 msr->r_msg = ERR_PTR(-E2BIG); pipelined_send()
586 msr->r_msg = NULL; pipelined_send()
587 msq->q_lrpid = task_pid_vnr(msr->r_tsk); pipelined_send()
589 wake_up_process(msr->r_tsk); pipelined_send()
597 msr->r_msg = msg; pipelined_send()
/linux-4.1.27/drivers/video/fbdev/i810/
H A Di810_gtf.c129 u8 msr = 0; i810fb_encode_registers() local
191 msr |= 1 << 6; i810fb_encode_registers()
193 msr |= 1 << 7; i810fb_encode_registers()
194 par->regs.msr = msr; i810fb_encode_registers()
H A Di810.h222 u8 msr; member in struct:mode_registers
242 u8 cr39, cr41, cr70, sr01, msr; member in struct:state_registers
/linux-4.1.27/drivers/ssb/
H A Ddriver_extif.c43 u8 save_mcr, msr = 0; serial_exists() local
48 msr = regs[UART_MSR] & (UART_MSR_DCD | UART_MSR_RI serial_exists()
52 return (msr == (UART_MSR_DCD | UART_MSR_CTS)); serial_exists()
/linux-4.1.27/drivers/misc/
H A Dcs5535-mfgpt.c50 uint32_t msr, mask, value, dummy; cs5535_mfgpt_toggle_event() local
69 msr = MSR_MFGPT_NR; cs5535_mfgpt_toggle_event()
74 msr = MSR_MFGPT_NR; cs5535_mfgpt_toggle_event()
79 msr = MSR_MFGPT_IRQ; cs5535_mfgpt_toggle_event()
87 rdmsr(msr, value, dummy); cs5535_mfgpt_toggle_event()
94 wrmsr(msr, value, dummy); cs5535_mfgpt_toggle_event()
/linux-4.1.27/drivers/ide/
H A Dcs5536.c29 * driver can be loaded with the "msr=1" parameter which forces using
38 #include <asm/msr.h>
305 module_param_named(msr, use_msr, int, 0644);
306 MODULE_PARM_DESC(msr, "Force using MSR to configure IDE function (Default: 0)");
/linux-4.1.27/arch/x86/platform/efi/
H A Defi_stub_64.S11 #include <asm/msr.h>
/linux-4.1.27/arch/x86/platform/olpc/
H A Dolpc-xo1-rtc.c17 #include <asm/msr.h>
/linux-4.1.27/arch/microblaze/include/uapi/asm/
H A Dptrace.h50 microblaze_reg_t msr; member in struct:pt_regs
/linux-4.1.27/drivers/ata/
H A Dpata_cs5536.c29 * driver can be loaded with the "msr=1" parameter which forces using
43 #include <asm/msr.h>
45 module_param_named(msr, use_msr, int, 0644);
46 MODULE_PARM_DESC(msr, "Force using MSR to configure IDE function (Default: 0)");
/linux-4.1.27/arch/powerpc/kernel/vdso32/
H A Dsigtramp.S69 /* If msr bit 1<<25 is set, then VMX register REGNO is at offset REGNO*16
87 .uleb128 33*RSIZE; /* msr offset */ \
101 /* If msr bit 1<<25 is set, then VMX register REGNO is at offset REGNO*16
112 /* If msr bit 1<<25 is set, then VMX register REGNO is at offset OFS of
/linux-4.1.27/arch/powerpc/kernel/vdso64/
H A Dsigtramp.S68 /* If msr bit 1<<25 is set, then VMX register REGNO is at offset REGNO*16
86 .uleb128 33*RSIZE; /* msr offset */ \
101 /* If msr bit 1<<25 is set, then VMX register REGNO is at offset REGNO*16
112 /* If msr bit 1<<25 is set, then VMX register REGNO is at offset OFS of
/linux-4.1.27/tools/perf/arch/powerpc/util/
H A Ddwarf-regs.c64 REG_DWARFNUM_NAME("%msr", 66),
/linux-4.1.27/drivers/net/hamradio/
H A Dbaycom_ser_fdx.c291 unsigned char iir, msr; ser12_interrupt() local
301 msr = inb(MSR(dev->base_addr)); ser12_interrupt()
303 if ((msr & 8) && bc->opt_dcd) ser12_interrupt()
304 hdlcdrv_setdcd(&bc->hdrv, !((msr ^ bc->opt_dcd) & 0x80)); ser12_interrupt()
335 msr = inb(MSR(dev->base_addr)); ser12_interrupt()
337 if ((msr & 8) && bc->opt_dcd) ser12_interrupt()
338 hdlcdrv_setdcd(&bc->hdrv, !((msr ^ bc->opt_dcd) & 0x80)); ser12_interrupt()
343 ser12_rx(dev, bc, &tv, msr & 0x10); /* CTS */ ser12_interrupt()
/linux-4.1.27/arch/x86/kernel/cpu/mtrr/
H A Dgeneric.c16 #include <asm/msr.h>
454 void mtrr_wrmsr(unsigned msr, unsigned a, unsigned b) mtrr_wrmsr() argument
456 if (wrmsr_safe(msr, a, b) < 0) { mtrr_wrmsr()
459 smp_processor_id(), msr, a, b); mtrr_wrmsr()
466 * @msr: MSR address of the MTTR which should be checked and updated
470 static void set_fixed_range(int msr, bool *changed, unsigned int *msrwords) set_fixed_range() argument
474 rdmsr(msr, lo, hi); set_fixed_range()
477 mtrr_wrmsr(msr, msrwords[0], msrwords[1]); set_fixed_range()
H A Dcentaur.c5 #include <asm/msr.h>
H A Damd.c4 #include <asm/msr.h>
H A Dcyrix.c8 #include <asm/msr.h>
/linux-4.1.27/arch/arm/mm/
H A Dproc-feroceon.S264 msr cpsr_c, r3 @ disable interrupts
267 msr cpsr_c, r2 @ restore interrupts
310 msr cpsr_c, r3 @ disable interrupts
313 msr cpsr_c, r2 @ restore interrupts
342 msr cpsr_c, r3 @ disable interrupts
345 msr cpsr_c, r2 @ restore interrupts
373 msr cpsr_c, r3 @ disable interrupts
376 msr cpsr_c, r2 @ restore interrupts
/linux-4.1.27/tools/power/cpupower/utils/
H A Dcpupower.c206 stat("/dev/cpu/0/msr", &statbuf) != 0) { main()
207 if (system("modprobe msr") == -1) main()
/linux-4.1.27/arch/x86/boot/
H A Dcpuflags.c6 #include <asm/msr-index.h>
/linux-4.1.27/arch/m68k/include/asm/
H A Dbvme6000hw.h50 pad_a[3], msr, member in struct:__anon1786
/linux-4.1.27/arch/arm/mach-s3c64xx/
H A Dsleep.S43 msr cpsr_c, #PSR_I_BIT | PSR_F_BIT | SVC_MODE
/linux-4.1.27/arch/arm/probes/kprobes/
H A Dactions-arm.c181 "msr cpsr_fs, %[cpsr] \n\t" emulate_rd12rn16rm0rs8_rwflags()
211 "msr cpsr_fs, %[cpsr] \n\t" emulate_rd12rn16rm0_rwflags_nopc()
241 "msr cpsr_fs, %[cpsr] \n\t" emulate_rd16rn12rm0rs8_rwflags_nopc()
291 "msr cpsr_fs, %[cpsr] \n\t" emulate_rdlo12rdhi16rn0rm8_rwflags_nopc()
/linux-4.1.27/arch/x86/pci/
H A Dmmconfig-shared.c190 u64 base, msr; pci_mmcfg_amd_fam10h() local
201 msr = high; pci_mmcfg_amd_fam10h()
202 msr <<= 32; pci_mmcfg_amd_fam10h()
203 msr |= low; pci_mmcfg_amd_fam10h()
206 if (!(msr & FAM10H_MMIO_CONF_ENABLE)) pci_mmcfg_amd_fam10h()
209 base = msr & (FAM10H_MMIO_CONF_BASE_MASK<<FAM10H_MMIO_CONF_BASE_SHIFT); pci_mmcfg_amd_fam10h()
211 busnbits = (msr >> FAM10H_MMIO_CONF_BUSRANGE_SHIFT) & pci_mmcfg_amd_fam10h()
/linux-4.1.27/net/mac80211/
H A Drc80211_minstrel.c340 struct minstrel_rate *msr, *mr; minstrel_get_rate() local
396 msr = &mi->r[ndx]; minstrel_get_rate()
404 msr->perfect_tx_time > mr->perfect_tx_time && minstrel_get_rate()
405 msr->stats.sample_skipped < 20) { minstrel_get_rate()
416 if (!msr->sample_limit) minstrel_get_rate()
420 if (msr->sample_limit > 0) minstrel_get_rate()
421 msr->sample_limit--; minstrel_get_rate()
/linux-4.1.27/arch/powerpc/xmon/
H A Dxmon.c398 return ((regs->msr & MSR_RI) == 0); unrecoverable_excp()
453 if ((regs->msr & (MSR_IR|MSR_PR|MSR_64BIT)) == (MSR_IR|MSR_64BIT)) xmon_core()
577 if (regs->msr & MSR_DE) { xmon_core()
585 if ((regs->msr & (MSR_IR|MSR_PR|MSR_64BIT)) == (MSR_IR|MSR_64BIT)) { xmon_core()
635 if ((regs->msr & (MSR_IR|MSR_PR|MSR_64BIT)) != (MSR_IR|MSR_64BIT)) xmon_bpt()
666 if ((regs->msr & (MSR_IR|MSR_PR|MSR_64BIT)) != (MSR_IR|MSR_64BIT)) xmon_break_match()
676 if ((regs->msr & (MSR_IR|MSR_PR|MSR_64BIT)) != (MSR_IR|MSR_64BIT)) xmon_iabr_match()
701 if ((regs->msr & (MSR_IR|MSR_PR|MSR_64BIT)) == (MSR_IR|MSR_64BIT)) { xmon_fault_handler()
974 regs->msr |= MSR_DE; do_step()
989 if ((regs->msr & (MSR_64BIT|MSR_PR|MSR_IR)) == (MSR_64BIT|MSR_IR)) { do_step()
1006 regs->msr |= MSR_SE; do_step()
1454 if (regs->msr & MSR_PR) print_bug_trap()
1491 printf(" msr: %lx\n", fp->msr); excprint()
1564 printf("msr = "REG" cr = %.8lx\n", fp->msr, fp->ccr); prregs()
1691 printf("msr = "REG" sprg0= "REG"\n", super_regs()
2501 "pc", "msr", "or3", "ctr", "lr", "xer", "ccr",
/linux-4.1.27/arch/x86/kernel/acpi/
H A Dwakeup_64.S6 #include <asm/msr.h>
/linux-4.1.27/drivers/bluetooth/
H A Ddtl1_cs.c292 unsigned char msr; dtl1_interrupt() local
339 msr = inb(iobase + UART_MSR); dtl1_interrupt()
341 if (info->ri_latch ^ (msr & UART_MSR_RI)) { dtl1_interrupt()
342 info->ri_latch = msr & UART_MSR_RI; dtl1_interrupt()
/linux-4.1.27/arch/powerpc/platforms/cell/
H A Dpervasive.c89 switch (regs->msr & SRR1_WAKEMASK) { cbe_system_reset_exception()
/linux-4.1.27/arch/powerpc/include/uapi/asm/
H A Dkvm_para.h47 __u64 msr; member in struct:kvm_vcpu_arch_shared
/linux-4.1.27/arch/frv/kernel/
H A Dasm-offsets.c74 DEF_FREG(__FPMEDIA_MSR0, f.msr[0]); foo()
/linux-4.1.27/drivers/isdn/i4l/
H A Disdn_tty.c301 info->msr &= ~UART_MSR_CTS; isdn_tty_tint()
446 info->msr &= ~UART_MSR_CTS; isdn_tty_senddown()
727 if ((info->msr & UART_MSR_RI) && isdn_tty_modem_hup()
730 info->msr &= ~(UART_MSR_DCD | UART_MSR_RI); isdn_tty_modem_hup()
1074 info->msr |= (UART_MSR_DSR | UART_MSR_CTS); isdn_tty_startup()
1092 info->msr &= ~UART_MSR_RI; isdn_tty_shutdown()
1159 info->msr |= UART_MSR_CTS; isdn_tty_write()
1201 info->msr |= UART_MSR_CTS; isdn_tty_write()
1363 status = info->msr; isdn_tty_tiocmget()
1757 return info->msr & UART_MSR_DCD; isdn_tty_carrier_raised()
2006 info->msr |= UART_MSR_RI; isdn_tty_find_icall()
2047 info->msr |= UART_MSR_CTS; isdn_tty_stat_callback()
2123 info->msr |= UART_MSR_DCD; isdn_tty_stat_callback()
2742 if (info->msr & UART_MSR_DCD) isdn_tty_cmd_ATand()
2979 if (info->msr & UART_MSR_RI) { isdn_tty_cmd_ATA()
2984 info->msr &= ~UART_MSR_RI; isdn_tty_cmd_ATA()
3424 if (info->msr & UART_MSR_DCD) isdn_tty_parse_at()
3426 if (info->msr & UART_MSR_RI) { isdn_tty_parse_at()
3503 if (info->msr & UART_MSR_DCD) isdn_tty_parse_at()
3546 if (info->msr & UART_MSR_DCD) { isdn_tty_parse_at()
3727 if (info->msr & UART_MSR_RI) { isdn_tty_modem_ring()
/linux-4.1.27/arch/mips/include/asm/mach-loongson/cs5536/
H A Dcs5536.h13 extern void _rdmsr(u32 msr, u32 *hi, u32 *lo);
14 extern void _wrmsr(u32 msr, u32 hi, u32 lo);
/linux-4.1.27/drivers/staging/rtl8192e/rtl8192e/
H A Dr8192E_dev.c63 u8 msr; rtl8192e_update_msr() local
66 msr = read_nic_byte(dev, MSR); rtl8192e_update_msr()
67 msr &= ~MSR_LINK_MASK; rtl8192e_update_msr()
72 msr |= (MSR_LINK_MANAGED << MSR_LINK_SHIFT); rtl8192e_update_msr()
74 msr |= (MSR_LINK_NONE << MSR_LINK_SHIFT); rtl8192e_update_msr()
79 msr |= (MSR_LINK_ADHOC << MSR_LINK_SHIFT); rtl8192e_update_msr()
81 msr |= (MSR_LINK_NONE << MSR_LINK_SHIFT); rtl8192e_update_msr()
85 msr |= (MSR_LINK_MASTER << MSR_LINK_SHIFT); rtl8192e_update_msr()
87 msr |= (MSR_LINK_NONE << MSR_LINK_SHIFT); rtl8192e_update_msr()
93 write_nic_byte(dev, MSR, msr); rtl8192e_update_msr()
/linux-4.1.27/arch/x86/xen/
H A Denlighten.c64 #include <asm/msr-index.h>
1033 static u64 xen_read_msr_safe(unsigned int msr, int *err) xen_read_msr_safe() argument
1037 val = native_read_msr_safe(msr, err); xen_read_msr_safe()
1038 switch (msr) { xen_read_msr_safe()
1049 static int xen_write_msr_safe(unsigned int msr, unsigned low, unsigned high) xen_write_msr_safe() argument
1055 switch (msr) { xen_write_msr_safe()
1083 ret = native_write_msr_safe(msr, low, high); xen_write_msr_safe()
1761 uint32_t eax, ebx, ecx, edx, pages, msr, base; init_hvm_pv_info() local
1771 cpuid(base + 2, &pages, &msr, &ecx, &edx); init_hvm_pv_info()
1774 wrmsr_safe(msr, (u32)pfn, (u32)(pfn >> 32)); init_hvm_pv_info()
/linux-4.1.27/drivers/acpi/
H A Dprocessor_throttling.c720 u64 msr = 0; acpi_throttling_rdmsr() local
732 msr = (msr_high << 32) | msr_low; acpi_throttling_rdmsr()
733 *value = (u64) msr; acpi_throttling_rdmsr()
742 u64 msr; acpi_throttling_wrmsr() local
749 msr = value; acpi_throttling_wrmsr()
751 msr & 0xffffffff, msr >> 32); acpi_throttling_wrmsr()
/linux-4.1.27/arch/x86/kernel/cpu/microcode/
H A Dintel_lib.c32 #include <asm/msr.h>
/linux-4.1.27/arch/x86/oprofile/
H A Dop_model_ppro.c19 #include <asm/msr.h>
/linux-4.1.27/arch/x86/realmode/rm/
H A Dreboot.S5 #include <asm/msr-index.h>

Completed in 3341 milliseconds

12